EUNO.NEWS EUNO.NEWS
  • All (20931) +237
  • AI (3154) +13
  • DevOps (932) +6
  • Software (11018) +167
  • IT (5778) +50
  • Education (48)
  • Notice
  • All (20931) +237
    • AI (3154) +13
    • DevOps (932) +6
    • Software (11018) +167
    • IT (5778) +50
    • Education (48)
  • Notice
  • All (20931) +237
  • AI (3154) +13
  • DevOps (932) +6
  • Software (11018) +167
  • IT (5778) +50
  • Education (48)
  • Notice
Sources Tags Search
한국어 English 中文
  • 5 days ago · ai

    Flexora: Flexible Low-Rank Adaptation for Large Language Models

    Vấn đề hiện tại - Bối cảnh: Việc tinh chỉnh fine‑tuning các mô hình ngôn ngữ lớn LLM tốn rất nhiều tài nguyên. Phương pháp LoRA Low‑Rank Adaptation ra đời để g...

    #LLM #LoRA #parameter-efficient fine-tuning #low-rank adaptation #Flexora
  • 6 days ago · ai

    An introduction to AWS Bedrock

    The how, why, what and where of Amazon’s LLM access layer The post An introduction to AWS Bedrock appeared first on Towards Data Science....

    #AWS #Bedrock #LLM #Amazon #generative AI #foundation models #cloud AI
  • 6 days ago · ai

    DeepSeek’s conditional memory fixes silent LLM waste: GPU cycles lost to static lookups

    When an enterprise LLM retrieves a product name, technical specification, or standard contract clause, it's using expensive GPU computation designed for complex...

    #LLM #conditional memory #GPU efficiency #inference optimization #AI infrastructure #model serving
  • 6 days ago · ai

    How Large Language Models (LLMs) Actually Generate Text

    !Cover image for How Large Language Models LLMs Actually Generate Texthttps://media2.dev.to/dynamic/image/width=1000,height=420,fit=cover,gravity=auto,format=au...

    #large language models #LLM #text generation #next-token prediction #deep learning #AI fundamentals
  • 6 days ago · ai

    Conversation Memory Collapse: Why Excessive Context Weakens AI

    Every story begins with a small misunderstanding. A midsize company approached us to build an AI support agent. Their request was simple—AI should “remember eve...

    #LLM #context window #prompt engineering #AI chatbots #memory collapse
  • 6 days ago · ai

    Cowork: Claude Code for the rest of your work

    Article URL: https://claude.com/blog/cowork-research-preview Comments URL: https://news.ycombinator.com/item?id=46593022 Points: 71 Comments: 19...

    #Claude #AI coding assistant #productivity #LLM #code generation #Anthropic
  • 6 days ago · ai

    Why your LLM bill is exploding — and how semantic caching can cut it by 73%

    Our LLM API bill was growing 30% month-over-month. Traffic was increasing, but not that fast. When I analyzed our query logs, I found the real problem: Users as...

    #LLM #semantic caching #API cost reduction #prompt optimization #AI infrastructure
  • 1 week ago · ai

    The `/context` Command: X-Ray Vision for Your Tokens

    'Stop guessing where your tokens go. Start seeing the invisible tax on your context window. From: x.com/adocomplete

    #token management #context window #Claude #LLM #prompt engineering #AI tooling
  • 1 week ago · ai

    TimeCapsuleLLM: LLM trained only on data from 1800-1875

    Article URL: https://github.com/haykgrigo3/TimeCapsuleLLM Comments URL: https://news.ycombinator.com/item?id=46590280 Points: 107 Comments: 51...

    #LLM #historical data #TimeCapsuleLLM #AI research #training dataset
  • 1 week ago · ai

    Why 90% Accuracy in Text-to-SQL is 100% Useless

    The eternal promise of self-service analytics The post Why 90% Accuracy in Text-to-SQL is 100% Useless appeared first on Towards Data Science....

    #text-to-sql #natural-language-processing #SQL #accuracy-metrics #self-service-analytics #LLM #AI-evaluation
  • 1 week ago · ai

    Why Ontario Digital Service couldn't procure '98% safe' LLMs (15M Canadians)

    Article URL: https://rosetta-labs-erb.github.io/authority-boundary-ledger/ Comments URL: https://news.ycombinator.com/item?id=46589386 Points: 16 Comments: 2...

    #Ontario Digital Service #LLM #AI safety #procurement #government #Canada
  • 1 week ago · ai

    When Does Adding Fancy RAG Features Work?

    Looking at the performance of different pipelines The post When Does Adding Fancy RAG Features Work? appeared first on Towards Data Science....

    #retrieval-augmented-generation #RAG #LLM #prompt-engineering #pipeline-performance #NLP #AI-tools

Newer posts

Older posts
EUNO.NEWS
RSS GitHub © 2026