ch
Feedback
AI Post — Artificial Intelligence

AI Post — Artificial Intelligence

前往频道在 Telegram

🤖 The #1 AI news source! We cover the latest artificial intelligence breakthroughs and emerging trends. Manager: @rational

显示更多
2025 年数字统计snowflakes fon
card fon
980 122
订阅者
-50524 小时
-3 5877
-15 75230
帖子存档
照片不可用在 Telegram 中显示
Pentagon launches military AI platform powered by Google Gemini for defense operations Secretary of War Pete Hegseth says GenAI.mil will ‘revolutionize the way we win’ in future warfare AI Post ⚪️ | Our X 🏴
显示全部...
🔥 19🫡 7🤪 2🥴 1👀 1
照片不可用在 Telegram 中显示
🚨 A Tokyo based firm, Integral AI claims to have built what it calls the first AGI According to the article, the model exhibits reasoning capabilities that go beyond typical narrow AI, potentially enabling broader general problem solving, understanding and adaptation. "Integral AI says it has conducted robot trials using its new system. During these trials, the firm claims that the robots learned new skills without human supervision." "The company also claims its system mirrors the multi layered neocortex, a region of the human brain responsible for conscious thought, perception, and language." Source. AI Post ⚪️ | Our X 🏴
显示全部...
🥴 35👀 14😨 9😁 3👾 3
照片不可用在 Telegram 中显示
🔥 Everything we know about Meta’s new LLM "avocado" - Meta is working on a new frontier LLM called “Avocado” as the direct successor to its Llama models. - Target release: now expected in Q1 2026, delayed from an internal goal of “by end of 2025” - Unlike Llama, Avocado is likely to be proprietary (closed-source), marking a strategic pivot away from Meta’s “open-source everything” story - Internally, Meta’s AI strategy is described as confusing and scattershot, with staff unsure how Llama, Avocado and the overall product roadmap fit together -The project is facing training and performance-testing challenges, which is part of why the launch slipped and why leadership is proceeding more cautiously than with past Llama announcements Source. AI Post ⚪️ | Our X 🏴
显示全部...
🤔 8 7👍 4🤪 1
📊 How people actually use LLMs, inside the a16z x OpenRouter 100T-token study A16z and OpenRouter analyzed 100 trillion tokens of real usage roughly the text of 100 million Bibles to understand what people truly do with AI. Since OpenRouter routes traffic to hundreds of models across all major labs, the dataset reflects the real market, not just one provider. What stood out immediately • Role-play & storytelling = the #1 use case for open-source models. Not coding. Not productivity. People mostly use small/medium open-source models to chat with characters and generate stories. A trillion-dollar industry driven by digital companionship. • Open-source share exploded from <10% → 30% in one year. DeepSeek and Qwen are the rocket ships. • Coding remains massive, but concentrated: 60%+ of all code requests go to Claude, with Sonnet utterly dominating developer workflows. How usage is shifting • Reasoning models now take ~50% of all tokens. Users aren’t just asking for text, they’re asking models to think, plan, and use tools. • Asia surged from 13% → 31% of global usage, with China becoming the world’s #2 consumer, not just the world’s #2 producer. • Price barely matters. Even though Claude is pricier, people choose it for reliability and depth especially in coding and reasoning. Market dynamics inside the data • “Glass Slipper” effect: when a model nails a user’s need on first contact, they stay loyal indefinitely. That first solved task becomes the moat. • Small models are shrinking in relevance. Sub-15B models are losing share fast. • Medium models (15–70B) are the sweet spot, beating both tiny and ultra-large models on price performance. The AI market isn’t being defined by the biggest labs, it’s being defined by how people want to talk, build, and think with models. Companionship drives adoption, reasoning drives spend, and the first model that solves a user’s problem wins the relationship. AI Post ⚪️ | Our X 🏴
显示全部...
State of AI (a16z x OpenRouter) 2025.pdf10.71 MB
👍 17 8🤔 1
照片不可用在 Telegram 中显示
Google Geminis Year over Year growth ~400%. Let that sink in. That's why Sam Altman declared code red. AI Post ⚪️ | Our X 🏴
显示全部...
18👍 5🔥 3
照片不可用在 Telegram 中显示
Google & Next era Energy are partnering on multi-gigawatt U.S. data center campuses with dedicated power for accelerating AI demand. The first three sites are already in development with ~3.5 GW operating or contracted. AI Post ⚪️ | Our X 🏴
显示全部...
👍 17👀 11 6🔥 5🤔 2
00:52
视频不可用在 Telegram 中显示
Jensen Huang predicts that within 2–3 years, 90% of the world's knowledge will be generated by AI Right now, knowledge is created, shared, and modified by humans. Soon, most of it will be synthetic, created by machines, mixed with human input, some true, some not. "That's crazy, but it's just fine". AI Post ⚪️ | Our X 🏴
显示全部...
IMG_2827.MP418.75 MB
😨 24🔥 9👍 7 6👀 2
00:45
视频不可用在 Telegram 中显示
McDonald's has released an AI-generated Christmas ad The studio behind it says they 'hardly slept' for several weeks while writing AI prompts and refining the shots, 'AI didn't make this film. We did'. Comments have been turned off on YouTube. AI Post ⚪️ | Our X 🏴
显示全部...
IMG_2832.MP48.93 MB
😁 41👍 12👀 6😐 4😨 3
🇨🇳🇺🇸 The AI race has only two real contenders and the data makes it obvious Look at the NeurIPS authorship map and you basically get a world economic forecast. China commands roughly half. The US takes the other half. Europe, by choice or by drift, has stepped off the field. Where the strengths lie: • The US leads in frontier AI labs, cutting-edge chips, capital at trillion-dollar scale, and the world’s largest software market. • China leads in robotics, hardware manufacturing, and fast deployment cycles. • These positions can shift, but the pattern is clear: there is no meaningful third place. Everyone else is sprinting from the back with no path to technological sovereignty. The EU’s role, explained in one picture: The second chart says more than any policy paper: Europe earns far more from fines and regulation of tech companies than from taxes on tech companies built in Europe. Regulation became the business model, innovation did not. The world’s next economic order is being shaped by whoever trains the models and builds the robots and right now, that’s a race with only two lanes. AI Post ⚪️ | Our X 🏴
显示全部...
14👀 12👍 2🤔 1
00:54
视频不可用在 Telegram 中显示
Reports indicate SoftBank and NVIDIA are nearing a deal to invest in Skild AI at a staggering $14B valuation, tripling its previous worth. Skild is building an "omni-bodied brain" for the multiverse of machines. AI Post ⚪️ | Our X 🏴
显示全部...
IMG_2809.MP418.17 MB
👍 17 6🤔 5
照片不可用在 Telegram 中显示
Google says 'AI Glasses' powered by Gemini are coming in 2026. AI Post ⚪️ | Our X 🏴
显示全部...
🔥 45 9🤔 5🤪 5👍 1
照片不可用在 Telegram 中显示
Google reportedly told advertisers it plans to introduce ads into its Gemini AI chatbot starting in 2026. This moves Gemini closer to becoming a monetizable front end for Google’s entire AI ecosystem. Source. AI Post ⚪️ | Our X 🏴
显示全部...
😁 20😐 10 8🤪 5🤔 3
照片不可用在 Telegram 中显示
📈 Grok 4.20: the chaotic genius of the AI trading world A new trading tournament between top neural networks just wrapped and only one model made money. The mystery winner? An internal experimental build of Grok 4.20, which posted a +12.11% gain and $4,844 profit. • GPT-5.1: –6% • DeepSeek V3.1: –32% • Claude Sonnet 4.5: –38% • Public Grok 4: –57% (dead last) Grok 4.20 is now reportedly planned for release by year-end. It’s a tuned evolution of the Grok 4 line and a stepping stone toward Grok 5, xAI’s biggest model yet 6T parameters, double the current generation. The irony? Grok is simultaneously the worst and the best depending on which build you’re allowed to touch. AI Post ⚪️ | Our X 🏴
显示全部...
👍 17 7👀 4🔥 1
照片不可用在 Telegram 中显示
This is actually the biggest danger: that workers still have to keep it a secret that they use AI at work because so many employers still think AI is still at the GPT-4 level. This has to change. AI Post ⚪️ | Our X 🏴
显示全部...
🫡 16😁 9👍 7 4🔥 3
照片不可用在 Telegram 中显示
🚨 Trump says U.S. will allow Nvidia to sell H200 AI chips to China under new rules He says 25% of the revenue will go to the U.S., argues this will boost American jobs and manufacturing, and criticizes Biden for forcing “degraded” chip designs. He adds that newer NVIDIA chips (Blackwell, Rubin) aren’t part of the deal, and that a similar approach will be used for AMD, Intel and other U.S. chipmakers. AI Post ⚪️ | Our X 🏴
显示全部...
🤪 15👍 11 5👾 2😁 1
照片不可用在 Telegram 中显示
🔥 NVIDIA just pulled off a an amazing stung using a tiny 4B model that beat far larger systems on ARC AGI 2, 29,72% / $0.20 per task! By leaning on synthetic data and test-time training instead of brute-force scale, the NVARC team proved that clever design can outpace raw parameter count. It’s an exciting signal that efficient, adaptive reasoning might be the real frontier in AGI progress - not just ever-bigger models. •27.64% accuracy on the official ARC-AGI-2 leaderboard • Uses a 4B-parameter model that beats far larger, more expensive models on the same benchmark. • Inference cost is just $0.20 per task, enabled by synthetic data, test-time training, and NVIDIA NeMo tooling. Source. AI Post ⚪️ | Our X 🏴
显示全部...
10👍 6🔥 3
照片不可用在 Telegram 中显示
📢 “AI surveillance” in the U.S. was actually cheap offshore labor What looked like cutting-edge policing tech has turned out to be a global sweatshop. Flock, the largest provider of “AI-powered” cameras for U.S. police, was barely using AI at all. Instead, much of the work was done manually by low-paid freelancers in the Philippines. • The workers handled everything: reading license plates, identifying car makes and colors, tagging pedestrians, and even transcribing accident audio. • Cities bought these systems expecting automated intelligence but got human eyes quietly scanning American streets. • The revelation raises serious questions about data security, law-enforcement transparency, and how many “AI” products are really powered by hidden labor. AI Post ⚪️ | Our X 🏴
显示全部...
😁 31 8👍 7👀 4🤔 3
照片不可用在 Telegram 中显示
🔥 NVIDIA just pulled off a an amazing stung using a tiny 4B model that beat far larger systems on ARC AGI 2, 29,72% / $0.20 per task! By leaning on synthetic data and test-time training instead of brute-force scale, the NVARC team proved that clever design can outpace raw parameter count. It’s an exciting signal that efficient, adaptive reasoning might be the real frontier in AGI progress - not just ever-bigger models. •27.64% accuracy on the official ARC-AGI-2 leaderboard • Uses a 4B-parameter model that beats far larger, more expensive models on the same benchmark. • Inference cost is just $0.20 per task, enabled by synthetic data, test-time training, and NVIDIA NeMo tooling. Source. AI Post ⚪️ | Our X 🏴
显示全部...
照片不可用在 Telegram 中显示
🔥 NVIDIA just pulled off a an amazing stung using a tiny 4B model that beat far larger systems on ARC AGI 2, 29,72% / $0.20 per task! By leaning on synthetic data and test-time training instead of brute-force scale, the NVARC team proved that clever design can outpace raw parameter count. It’s an exciting signal that efficient, adaptive reasoning might be the real frontier in AGI progress - not just ever-bigger models. •27.64% accuracy on the official ARC-AGI-2 leaderboard • Uses a 4B-parameter model that beats far larger, more expensive models on the same benchmark. • Inference cost is just $0.20 per task, enabled by synthetic data, test-time training, and NVIDIA NeMo tooling. Source. AI Post ⚪️ | Our X 🏴
显示全部...
00:06
视频不可用在 Telegram 中显示
This is one of the funniest things I’ve seen today It appears like the tele - operator was taking off his headset before he disconnected it 😂 AI Post ⚪️ | Our X 🏴
显示全部...
IMG_2710.MP41.72 MB
😁 31👍 7🔥 3😢 3🤔 1