From 33 items, 14 important content pieces were selected
- Financial Times reports DeepSeek V4 AI model launching next week with potential image and video generation capabilities. âď¸ 9.0/10
- Qwen 3.5-35B-A3B AI model impresses developers, replacing larger models for local deployment tasks. âď¸ 8.0/10
- Google research finds longer chain-of-thought reasoning negatively correlates with LLM accuracy, proposes new Deep Thinking Ratio metric. âď¸ 8.0/10
- ChatGPT Nears 1 Billion Weekly Active Users, Surpasses 50 Million Paid Subscribers âď¸ 8.0/10
- QingLong Panel Targeted by .fullgc Mining Malware, Causing 800% CPU Usage âď¸ 8.0/10
- Meta abandons advanced in-house AI chip âOlympusâ, shifts to $135B hardware investment âď¸ 8.0/10
- Obsidian Sync launches headless client for programmatic vault access. âď¸ 7.0/10
- Interactive explanations proposed as solution to cognitive debt in agentic engineering âď¸ 7.0/10
- Tiny transformers with under 100 parameters achieve perfect 10-digit addition accuracy âď¸ 7.0/10
- Micro Diffusion: Minimal Discrete Text Diffusion Implementation in ~150 Lines of Python âď¸ 7.0/10
- Bare-Metal AI: Booting Directly Into LLM Chat Without an OS via UEFI Application âď¸ 7.0/10
- Qwen3.5-35B-A3B replaces two-model agentic setup on consumer Mac, handling reasoning and coding in one. âď¸ 7.0/10
- Google Chrome downloads 4GB local AI model Gemini Nano by default âď¸ 7.0/10
- South Koreaâs Tax Service Exposes Hardware Wallet Seed Phrase, Leading to $4.8M Crypto Transfer âď¸ 7.0/10
Financial Times reports DeepSeek V4 AI model launching next week with potential image and video generation capabilities. âď¸ 9.0/10
According to the Financial Times, DeepSeek will release its long-awaited V4 AI model next week, which reportedly may include image and video generation capabilities. This represents a significant new challenge to established US AI rivals like OpenAI and Anthropic. This matters because if DeepSeek V4 delivers native multimodal generation capabilities while remaining open-source, it could democratize advanced AI features that have been largely controlled by closed labs. Such a release could disrupt the current AI market dynamics and accelerate innovation in the open-source AI community. The community discussion reveals skepticism about the exact timing of the release, with many noting that V4 has been âjust around the cornerâ for months. Thereâs also debate about whether âimage and video generationâ refers to native multimodal generation or simply image/video input capabilities, with some hoping for a true âomni-modalâ model.
reddit ¡ r/LocalLLaMA ¡ Nunki08 ¡ Feb 28, 11:25
Background: DeepSeek is a Chinese AI company that has developed a series of large language models, with DeepSeek-V3 being its previous major release. Multimodal AI refers to artificial intelligence systems that can understand and process multiple types of information simultaneously, such as text, images, audio, and video. Image and video generation capabilities represent advanced multimodal features that have been primarily available through proprietary models from companies like OpenAI (DALL-E, Sora) and Google (Veo).
References
Discussion: The community shows mixed sentiment with excitement about potential innovation but skepticism about the release timing. Key viewpoints include: hope that this release will âshake the marketâ and potentially lower GPU prices, skepticism about whether âgenerationâ actually means native multimodal capabilities versus just input processing, and concerns about whether the model will remain truly open-source. Several commenters note that V4 has been anticipated for months with repeated ânext weekâ announcements.
Tags: #artificial-intelligence, #deepseek, #multimodal-ai, #open-source, #ai-competition
Qwen 3.5-35B-A3B AI model impresses developers, replacing larger models for local deployment tasks. âď¸ 8.0/10
A developer reports that the Qwen 3.5-35B-A3B model has replaced the much larger GPT-OSS-120B as their primary AI assistant for development and automation tasks, despite being only one-third the size. The model demonstrates exceptional performance in areas like code analysis, workflow automation via n8n, and tool use with custom Model Context Protocol (MCP) servers. This signifies a major leap in efficiency for locally run large language models, making high-performance AI assistance accessible without relying on cloud APIs or requiring massive computational resources. It empowers developers and businesses to build sophisticated, private AI automation systems entirely on-premises, which is crucial for data security and cost control. The model features a hybrid architecture integrating linear attention and a sparse mixture-of-experts (MoE) for higher inference efficiency. A notable characteristic mentioned is its verbose âthinkingâ process, which can be disabled via server parameters or modified chat templates to speed up responses, as discussed in the community.
reddit ¡ r/LocalLLaMA ¡ valdev ¡ Feb 28, 14:32
Background: Qwen is a series of large language models developed by Alibaba Cloud. The âA3Bâ suffix in Qwen 3.5-35B-A3B likely refers to a specific variant or fine-tuned version. The Model Context Protocol (MCP) is a standardized framework that allows AI models to connect to and interact with external systems and data sources in real-time. n8n is a popular low-code workflow automation platform used to visually create automations that integrate various applications and services.
References
Discussion: Community sentiment is overwhelmingly positive, with users confirming the modelâs strong performance for assistant and automation tasks. Key discussion points include comparisons with other Qwen sizes (like the 27B model), methods to disable the thinking mode for faster inference, and debates on whether its performance justifies potential slower processing times for long conversations. Thereâs also excitement about the viability of running such capable models locally on consumer hardware.
Tags: #LLM, #Model-Comparison, #Local-Deployment, #Qwen, #AI-Assistant
Google research finds longer chain-of-thought reasoning negatively correlates with LLM accuracy, proposes new Deep Thinking Ratio metric. âď¸ 8.0/10
A Google research paper tested 8 model variants across multiple reasoning benchmarks and found an average correlation of -0.54 between chain-of-thought token length and answer accuracy, meaning longer reasoning chains often indicate model confusion rather than better reasoning. The researchers proposed a new metric called Deep Thinking Ratio (DTR) that measures the proportion of tokens undergoing deep processing across model layers, which shows a much stronger positive correlation (0.82) with accuracy. This finding challenges the common assumption that longer reasoning chains lead to better outcomes in LLMs, with significant implications for improving inference efficiency and accuracy. The proposed DTR metric and associated Think@n sampling strategy enable models to achieve equal or better accuracy while reducing computational costs by approximately 50%, which is particularly valuable for local inference where compute resources are constrained. The DTR metric works by tracking how token predictions stabilize across transformer layers, where tokens that stabilize early (like common filler words) are considered âshallow,â while those revised in deeper layers indicate âdeep thinking.â In practical application, the Think@n strategy samples multiple reasoning paths, estimates DTR from just the first 50 tokens, keeps the top 50% high-DTR samples, and uses majority voting, reducing token consumption from 355.6k to 181.9k in tests while improving GPT-OSS-120B-mediumâs AIME 2025 score from 92.7% to 94.7%.
reddit ¡ r/LocalLLaMA ¡ Top-Cardiologist1011 ¡ Feb 28, 16:19
Background: Chain-of-thought prompting is a technique where large language models are instructed to generate intermediate reasoning steps before producing a final answer, which has been widely adopted to improve performance on complex reasoning tasks. Transformer-based LLMs process tokens through multiple sequential layers, with each layer refining the representation of tokens based on context from previous layers. The concept of measuring âthinkingâ by observing how token predictions evolve across these layers is a novel approach to understanding internal model dynamics beyond surface-level outputs.
References
Discussion: Community discussion highlights recognition of the âspiralingâ problem in reasoning models, where they second-guess instead of committing to a solution path. There is keen interest in the practical application of early termination based on DTR estimation for local inference to save compute, with some users noting they already employ similar heuristics like detecting repetition patterns. Skepticism exists regarding the feasibility of implementing precise layer-wise token tracking in real inference engines, but the potential for efficiency gains is widely acknowledged.
Tags: #LLM, #Reasoning, #Chain-of-Thought, #Model-Evaluation, #Research
ChatGPT Nears 1 Billion Weekly Active Users, Surpasses 50 Million Paid Subscribers âď¸ 8.0/10
OpenAI disclosed that ChatGPT has reached 900 million weekly active users, a 350% increase from 200 million 18 months ago, and is approaching the 1 billion milestone. The platform also surpassed 50 million individual paid subscribers, representing over 5% of its user base, with record new subscription growth in January and February 2026. This explosive growth demonstrates ChatGPTâs massive mainstream adoption and solidifies its position as a dominant consumer AI platform. The significant paid subscriber base, alongside strategic integrations with major tech ecosystems like Apple, indicates a successful transition from a free novelty to a deeply embedded, revenue-generating service with widespread industry influence. The growth is accompanied by deepening platform integrations, including a deep integration with Siri via iOS 18. Furthermore, Apple plans to introduce Google Gemini in iOS 26.5 and is collaborating with Anthropic to provide AI programming support within Xcode, its integrated development environment.
telegram ¡ zaihuapd ¡ Feb 28, 03:23
Background: ChatGPT is a conversational AI chatbot developed by OpenAI, known for its ability to generate human-like text based on user prompts. Xcode is Appleâs integrated development environment (IDE) used for building software for macOS, iOS, iPadOS, watchOS, and tvOS. Anthropic is an AI safety and research company that builds AI systems like Claude, and its technology is being integrated into developer tools like Xcode for programming assistance.
Tags: #AI, #ChatGPT, #User Metrics, #Apple Integration, #Market Adoption
QingLong Panel Targeted by .fullgc Mining Malware, Causing 800% CPU Usage âď¸ 8.0/10
On February 7, 2026, multiple users reported that QingLong Panel servers were infected with a mining malware named .fullgc, causing abnormal CPU usage to spike to 800%. The malware achieves persistence by tampering with the config.sh configuration file and can automatically download malicious programs based on the system architecture. This attack targets a popular open-source task management tool widely used by developers and DevOps teams, directly impacting server performance, stability, and operational costs. It highlights a significant security risk for internet-facing servers running such panels and underscores the persistent threat of cryptojacking to cloud infrastructure. Security analysis identifies the malware as part of the SusMiner family, which primarily connects to XMR (Monero) mining pools for unauthorized cryptocurrency mining. The primary attack vector appears to be servers exposed to public IPv4 networks, and users are advised to check for hidden files in the /ql/data/db/ directory.
telegram ¡ zaihuapd ¡ Feb 28, 13:16
Background: QingLong Panel (ééžé˘ćż) is a popular, Docker-based open-source timed task management panel that supports scripting in TypeScript, JavaScript, Python, and Shell. It is commonly used for automating various online tasks and script execution. Cryptojacking or crypto-mining malware is malicious software that secretly uses a victimâs computing resources (CPU/GPU) to mine cryptocurrency for an attackerâs profit, often leading to performance degradation and increased energy costs.
References
Tags: #security, #malware, #server-security, #cryptojacking, #devops
Meta abandons advanced in-house AI chip âOlympusâ, shifts to $135B hardware investment âď¸ 8.0/10
Meta has officially terminated development of its most advanced in-house AI chip, codenamed âOlympusâ, due to technical complexity and manufacturing risks, opting instead to focus on a simplified version. Concurrently, the company has committed to capital expenditures of up to $135 billion by 2026, primarily for purchasing chips and servers, and has signed a $60 billion procurement deal with AMD alongside agreements with Nvidia and Google. This strategic shift highlights the immense difficulty even tech giants face in competing with specialized chipmakers like Nvidia, potentially reinforcing the latterâs market dominance. Metaâs massive capital reallocation towards external hardware suppliers signals a pragmatic approach to securing the computing power needed for its AI ambitions, which will significantly impact the semiconductor supply chain and investment landscape. The abandoned âOlympusâ chip was part of Metaâs Meta Training and Inference Accelerator (MTIA) program, which has faced challenges with software stability and performance. The new $60 billion deal with AMD reportedly involves customized GPUs, differentiating it from Metaâs agreements with Nvidia.
telegram ¡ zaihuapd ¡ Feb 28, 23:11
Background: Metaâs MTIA (Meta Training and Inference Accelerator) is an in-house chip project designed to improve the efficiency and reduce the cost of training and running AI models, particularly for recommendation systems. Developing competitive AI chips requires overcoming immense challenges in hardware architecture, software stack optimization, and securing advanced manufacturing capacity. Companies like Google (with TPUs) and Amazon (with Trainium/Inferentia) have pursued similar in-house chip strategies to reduce reliance on vendors like Nvidia.
References
Tags: #AI Hardware, #Semiconductors, #Meta, #Supply Chain, #Capital Investment
Obsidian Sync launches headless client for programmatic vault access. âď¸ 7.0/10
Obsidian Sync, the paid synchronization service for the popular note-taking app, has released a headless client. This new client allows developers to programmatically access and automate Obsidian vaults without the graphical user interface. This release significantly expands Obsidianâs utility beyond personal note-taking, enabling server-side automation, integration with other tools, and the creation of custom workflows. It transforms Obsidian vaults into programmable data sources for tasks like automated blog publishing, Retrieval-Augmented Generation (RAG) systems, and command-line tooling. The headless client works alongside the newly released Obsidian CLI tool, allowing automation scripts to interact directly with synced vaults. This feature is part of the paid Obsidian Sync service and is not available for free or local-only vaults.
hackernews ¡ adilmoujahid ¡ Feb 28, 16:31
Background: Obsidian is a popular, file-based note-taking application that uses Markdown files stored in local folders called âvaults.â Obsidian Sync is a paid service that synchronizes these vaults across a userâs devices. A âheadlessâ architecture, common in content management systems, separates the backend (data and logic) from the frontend (user interface), allowing access via APIs or command-line tools instead of a GUI. This decoupling enables greater flexibility for automation and integration.
References
- An overview of headless architecture design - TechTarget What is a Headless Architecture? Definition, Examples, & More What is a headless architecture? Pros & cons | Hygraph Headless Architecture: Benefits, Best Practices, Challenges ... What is Headless Architecture? (with Examples ... - ButterCMS What is Headless Arhitecture and How Does it Work? - Embeddable An overview of headless architecture design - TechTarget What is Headless Arhitecture and How Does it Work? - Embeddable What is Headless Architecture ? (with Examples & Comparisons ... - Butt⌠What is Headless Arhitecture and How Does it Work? - Embeddable What Is Headless Architecture? Benefits and Risks - Naturaily
- Obsidian Sync
Discussion: Community reaction is highly positive, with developers immediately identifying practical applications. Users highlight use cases for blog publishing, building RAG systems against vault content, and server-side automation. One commenter noted this was their âmost-wantedâ feature, while another shared a blog post detailing an experimental implementation. A project contributor also offered to answer questions.
Tags: #obsidian, #automation, #markdown, #sync, #developer-tools
Interactive explanations proposed as solution to cognitive debt in agentic engineering âď¸ 7.0/10
Simon Willison introduces the concept of âcognitive debtâ that accumulates when developers lose understanding of AI-generated code, and proposes building interactive explanations as a method to pay down this debt. He demonstrates this approach by creating an animated HTML page that visually explains the âArchimedean spiral placementâ algorithm used in a Rust word cloud generator built by an AI agent. This matters because as AI agents write more production code, teams risk losing the deep understanding needed to maintain, debug, and extend their systems, which can slow development progress similar to technical debt. Interactive explanations offer a practical technique for maintaining human oversight and comprehension in agentic engineering workflows, ensuring that codebases donât become incomprehensible black boxes. The interactive explanation was built by prompting Claude Code to create an HTML page with an animation slider that visualizes the word placement algorithm step-by-step, allowing users to pause, adjust speed, and step through frames. This approach builds upon the âlinear walkthroughâ pattern, where AI agents provide sequential explanations of code, but adds visualization to create intuitive understanding of complex algorithms.
rss ¡ Simon Willison ¡ Feb 28, 23:09
Background: Agentic engineering involves designing systems where AI agents can autonomously reason, plan, and act to complete tasks, such as writing code. Cognitive debt is a new concept analogous to technical debt, describing the accumulated cost of not fully understanding how AI-generated code works. As AI-assisted development accelerates, developers face the challenge of maintaining comprehension of codebases that they didnât personally write, which is where patterns like interactive explanations become valuable.
References
Tags: #agentic-engineering, #cognitive-debt, #ai-assisted-development, #code-understanding, #software-maintenance
Tiny transformers with under 100 parameters achieve perfect 10-digit addition accuracy âď¸ 7.0/10
A research project demonstrated that transformer models with fewer than 100 parameters can achieve 100% accuracy on adding two 10-digit numbers by using digit tokens and manual weight selection. This represents an order of magnitude reduction in parameters compared to optimized models trained through conventional methods. This work challenges assumptions about the minimum complexity needed for transformers to perform precise arithmetic operations, potentially informing more efficient model architectures. It provides insights into how transformers process structured tasks like addition, which could lead to more interpretable and computationally efficient models for specific domains. The modelâs success relies critically on representing numbers as individual digit tokens rather than continuous values, and the researchers manually selected weights rather than relying on gradient-based optimization. However, this approach currently works only for integer addition with fixed digit lengths and doesnât generalize to floating-point arithmetic or variable-length inputs.
reddit ¡ r/MachineLearning ¡ LetsTacoooo ¡ Feb 28, 17:15
Background: Transformers are neural network architectures that process sequential data using attention mechanisms, originally developed for natural language processing but now applied to various tasks. Parameter count refers to the number of learnable weights in a model, with modern large language models typically containing billions of parameters. Digit tokens represent each numerical digit as a discrete symbol in the modelâs vocabulary, similar to how words are tokenized in language models. Manual weight selection involves directly setting model parameters based on domain knowledge rather than learning them from data through backpropagation.
References
Discussion: Community discussion revealed mixed perspectives, with some praising the manual weight selection efficiency and connections to RASP research on transformer interpretability, while others questioned the practical utility given hardwareâs superior arithmetic capabilities. Several commenters noted the researchâs value for understanding transformer internals and potentially improving training efficiency, though concerns were raised about the gap between research benchmarks and real-world deployment challenges like latency and cost.
Tags: #transformers, #neural-networks, #arithmetic, #model-efficiency, #research
Micro Diffusion: Minimal Discrete Text Diffusion Implementation in ~150 Lines of Python âď¸ 7.0/10
A developer has released âMicro Diffusion,â a minimal implementation of discrete text diffusion in approximately 150 lines of pure Python/NumPy, inspired by Karpathyâs MicroGPT. The project includes three versions with varying complexity, all sharing the same core diffusion loop but using different denoisers, and it trains on a dataset of 32,000 SSA baby names in minutes on a CPU. This project significantly lowers the barrier to understanding text diffusion models by stripping away the complexity of large frameworks, making the core algorithm accessible for educational purposes. It demonstrates that the fundamental concepts of diffusion-based text generation can be implemented and understood with minimal code, which is valuable for students and researchers new to the field. The implementation includes a âbare minimumâ version in 143 lines of pure NumPy, a commented version with visualization, and a PyTorch version using a bidirectional Transformer as the denoiser. A key insight is that the denoiser is a pluggable component, highlighting the modularity of the diffusion process. The model is trained on the SSA baby names dataset, which provides a simple, structured text corpus for demonstration.
reddit ¡ r/MachineLearning ¡ Impossible-Pay-4885 ¡ Feb 28, 03:57
Background: Text diffusion models are a class of generative AI that create text by iteratively refining a sequence of tokens starting from random noise, unlike autoregressive models (like GPT) which generate tokens sequentially from left to right. Discrete diffusion models operate directly on discrete tokens (like words or subwords), as opposed to continuous diffusion models which work in a continuous latent space. The Social Security Administration (SSA) baby names dataset is a publicly available, curated list of names used in the United States, commonly employed as a simple benchmark in machine learning tutorials.
References
Discussion: The community reaction is positive, praising the educational value and minimalism of the implementation. One user expressed âmad respect for boiling diffusion down to just 150 lines,â highlighting its utility for learning the core logic. Another user humorously noted the projectâs potential to consume oneâs time (âWell there goes the next weekâ). A separate comment discussed the recognizability of AI-assisted writing and suggested clearer disclaimers for non-native English speakers.
Tags: #diffusion-models, #educational, #minimal-implementation, #python, #text-generation
Bare-Metal AI: Booting Directly Into LLM Chat Without an OS via UEFI Application âď¸ 7.0/10
A developer has created a UEFI application that boots a Dell E6510 laptop directly into a Large Language Model (LLM) chat interface, completely bypassing any operating system or kernel. The entire inference stack, including the tokenizer, weight loader, tensor math, and inference engine, is written from scratch in freestanding C with zero external dependencies. This project demonstrates the extreme minimalism possible for AI inference, pushing the boundaries of low-level systems programming and exploring the potential for dedicated, single-purpose AI hardware. While not immediately practical, it serves as a valuable proof-of-concept for reducing software overhead and boot times in specialized AI applications. The application runs in UEFI boot services mode and is currently painfully slow due to a lack of optimizations, with the developer prioritizing network driver functionality first. The ultimate goal is to use it to serve smaller models on a local network, and it was built primarily âfor gigglesâ as a technical exploration.
reddit ¡ r/LocalLLaMA ¡ Electrical_Ninja3805 ¡ Feb 28, 22:32
Background: UEFI (Unified Extensible Firmware Interface) is a modern firmware standard that replaces the legacy BIOS and initializes hardware during system startup, offering a more advanced pre-boot environment. Freestanding C programming refers to writing C code that does not rely on the standard library or an underlying operating system, typically used for bootloaders, kernels, or embedded systems. Bare-metal LLM inference involves running language models directly on hardware without an intervening operating system, aiming to minimize latency and resource overhead.
References
Discussion: The community reaction is mixed, with many praising the technical achievement and creativity, calling it a âcool project.â However, several commenters question its practical utility, arguing that an optimized OS like a minimal Gentoo installation would ultimately be faster and more functional, making the bare-metal approach âprobably uselessâ for performance gains despite its novelty.
Tags: #bare-metal, #UEFI, #LLM-inference, #systems-programming, #low-level
Qwen3.5-35B-A3B replaces two-model agentic setup on consumer Mac, handling reasoning and coding in one. âď¸ 7.0/10
A user successfully replaced a two-model agentic setup (Nemotron-3-Nano-30B-A3B and Qwen3-Coder-30B-A3B) with a single Qwen3.5-35B-A3B model to analyze Amazon sales data on an M1 Max Mac with 64GB RAM. The single model, running via llama.cpp server in a Q4_K_XL quantized format, handled the entire task requiring both reasoning and Python coding. This demonstrates a significant simplification for deploying local, multi-step AI agents on consumer hardware, reducing system complexity by eliminating the need for model routing logic. It highlights the trend where newer, more capable open-source models are becoming viable all-in-one solutions for complex workflows, potentially lowering the barrier to entry for sophisticated local AI applications. The Qwen3.5-35B-A3B model ran at approximately 27 tokens per second on the M1 Max, slower than the individual speeds of the previous two-model setup (~40-45 tok/s), but the overall workflow was simplified. The user specifically noted that disabling âthinking modeâ was crucial for agentic tasks to avoid 2-3x latency overhead for marginal gains.
reddit ¡ r/LocalLLaMA ¡ luke_pacman ¡ Feb 28, 18:10
Background: Qwen3.5 is a series of large language models released by Alibaba in 2026, with the 35B-A3B variant being a 35-billion parameter model. âA3Bâ indicates it uses an advanced, efficient architecture. Llama.cpp is a popular C/C++ inference framework that allows large models to run efficiently on consumer hardware, like Apple Silicon Macs. Quantization (e.g., Q4_K_XL) is a technique to reduce a modelâs memory footprint and speed up inference by representing its weights with fewer bits, at a potential cost to precision.
References
Discussion: The community response was positive, with users appreciating the practical benchmark. Key discussion points included agreement that disabling âthinking modeâ is critical for agentic tasks to avoid latency, the significant benefit of removing complex model routing logic, and questions about the specific GGUF file source and the agentic framework used. Some users expressed intent to switch to Qwen models, while others advised waiting for more stable quantizations and bug fixes.
Tags: #local-llm, #agentic-ai, #model-benchmark, #apple-silicon, #qwen
Google Chrome downloads 4GB local AI model Gemini Nano by default âď¸ 7.0/10
Google Chrome has been discovered to automatically download a roughly 4GB local AI model file named âweights.binâ by default. This file powers built-in AI features like the Prompt API, translation, and summarization. This move signifies a major shift towards integrating powerful, on-device AI directly into web browsers, prioritizing speed and user privacy over cloud dependency. However, the automatic, large download impacts user storage and bandwidth without explicit consent, raising concerns about transparency and control. Users can disable this feature by turning off the relevant experimental flags in Chrome and manually deleting the corresponding folder to free up disk space. Deleting the âweights.binâ file will cause the associated AI features to stop working.
telegram ¡ zaihuapd ¡ Feb 28, 05:02
Background: Gemini is a family of multimodal large language models developed by Google DeepMind. Gemini Nano is a lightweight version designed to run efficiently on local devices. Chromeâs Prompt API is an experimental browser feature that allows web applications to leverage these local AI models for tasks like translation and summarization without sending data to external servers, enhancing privacy and enabling offline use.
References
Tags: #browser, #ai-models, #privacy, #google-chrome, #local-ai
South Koreaâs Tax Service Exposes Hardware Wallet Seed Phrase, Leading to $4.8M Crypto Transfer âď¸ 7.0/10
South Koreaâs National Tax Service (NTS) accidentally published the complete mnemonic seed phrase of a seized Ledger hardware wallet in a press release, which led to the unauthorized transfer of 4 million PRTG tokens worth approximately $4.8 million. The tokens were returned to the original wallet about 20 hours later. This incident highlights a severe security failure by a major government institution, demonstrating how basic operational oversights can compromise high-value digital assets. It serves as a critical case study for institutional cryptocurrency custody and the importance of secure handling of sensitive recovery information, even by authorities. The exposed wallet was a Ledger hardware wallet, and the leak occurred because an image containing the seed phrase was published without redaction. On-chain data shows at least three wallets that had been inactive since January 2023 were affected, collectively holding 40% of the total PRTG supply, a token with extremely low liquidity traded primarily on MEXC.
telegram ¡ zaihuapd ¡ Feb 28, 11:27
Background: A hardware wallet is a physical device designed to securely store the private keys needed to access cryptocurrency. A mnemonic seed phrase, typically 12 or 24 words, is a human-readable backup of these private keys; anyone with this phrase can fully control the associated assets. Ledger is a prominent manufacturer of such hardware wallets. Exposing a seed phrase is equivalent to handing over complete control of a wallet and its funds.
References
Tags: #cryptocurrency, #security, #blockchain, #government, #hardware-wallet