1. Bluesky Feeds /
  2. Doll /
  3. AI Knowers

Posts from pro-AI posters and AI experts curated in the "AI Knowers" starter pack. Filters for AI-relevant content from this handpicked group of thinkers.

Feed on Bluesky

Feeds Stats

  • 💙 Liked by 26 users
  • 📅 Updated 23 days ago
  • ⚙️ Provider attie.ai
  • 📈 In the last 30 days, there were 5 posts about this feed. These posts got a total of 54 likes and had 2 reposts.

AI Knowers Likes over time

Like count prediction
The feed AI Knowers gains approximately 11 likes per month.

Feed Preview for AI Knowers

Liz Fong-Jones (方禮真)
@lizthegrey.com
42 minutes ago
it turns out to be less effort to just vibe an MCP server than to download someone else's and hope it's not compromised. *especially* in light of the Vercel google docs API compromise.
5
1
36
Tim Kellogg
@timkellogg.me
about 1 hour ago
Kimi 2.6: Hangs with the best * on par with Opus 4.6 & GPT-5.4 5.4 xhigh * long horizon coding tasks www.kimi.com/blog/kimi-k2-6
A bar chart titled "percentage (%)" comparing the performance of four AI models across various benchmarks. The models are: **Kimi K2.6** (blue), **GPT-5.4 (xhigh)** (white), **Claude Opus 4.6 (max effort)** (light gray), and **Gemini 3.1 Pro (thinking high)** (white).
The benchmarks are divided into three categories:
### General Agents
 * **Humanity’s Last Exam (Full) w/ tools:** Kimi K2.6 leads with 54.0.
 * **BrowseComp:** Gemini 3.1 Pro leads with 85.9; Kimi K2.6 follows at 83.2.
 * **DeepSearchQA (f1-score):** Kimi K2.6 leads significantly with 92.5.
 * **Toolathlon:** GPT-5.4 leads with 54.6; Kimi K2.6 is at 50.0.
 * **OSWorld-Verified:** GPT-5.4 leads with 75.0; Kimi K2.6 is at 73.1. (Claude is 72.7, Gemini is missing data).
### Coding
 * **Terminal-Bench 2.0 (Terminus-2):** Gemini 3.1 Pro leads with 68.5; Kimi K2.6 is at 66.7.
 * **SWE-Bench Pro:** Kimi K2.6 leads with 58.8.
 * **SWE-bench Multilingual:** Claude Opus 4.6 leads with 77.8; Kimi K2.6 is at 76.7. (Gemini is 76.9, GPT is missing data).
### Visual Agents
 * **MathVision w/ python:** GPT-5.4 leads with 96.1; Kimi K2.6 is at 93.2.
 * **V* w/ python:** GPT-5.4 leads with 98.4; Kimi K2.6 and Gemini 3.1 Pro are tied at 96.9.
1
2
24
Eris (prev. Isolyth)
@isolyth.dev
about 2 hours ago
New Kimi!!!! It's open weight immediately and it scores really really good on the benchmarks!!! It's gotten even better on long horizon work, seemingly a major focus for Moonshot www.kimi.com/blog/kimi-k2-6
2
1
25
SE Gyges
@segyges.bsky.social
about 2 hours ago
so now that uv lets you just import things by git url is there a purpose to having pypi really
10
0
30
LittleKuriboh
@littlekuriboh.bsky.social
about 3 hours ago
“miHoYo, with the Shanghai Jiao Tong University Student Innovation Center, will host a talk session on upgrading game industrialization pipelines through AI… … including AI agents, AI-driven gameplay, AI-powered NPCs, and AIGC (AI-generated content), highlighting real-world implementation.”
7
51
94
Eris (prev. Isolyth)
@isolyth.dev
about 4 hours ago
Well that sucks
2
1
27
Tim Kellogg
@timkellogg.me
about 4 hours ago
inside traders announce that GPT-5.5 “Spud” will be released on Thursday
That Polymarket chart is looking pretty spicy, especially with that massive spike for **April 23rd**. A 75% "probability" according to the bettors is significant—it suggests they are pricing in some very specific rumors or a scheduled event.
Since today is the 20th, we’re only three days out from that peak. If those odds hold true, the "AI spring" is about to get a lot more crowded.
### Why April 23rd?
The sharp vertical move on the blue line usually happens when:
 1. **Leaked Benchmarks:** "Leaked" results showing massive gains in reasoning or coding often hit X (Twitter) or GitHub around this time.
 2. **Conference Alignment:** OpenAI has a history of "coincidental" timing when competitors (like Google or Anthropic) have major announcements.
 3. **The "Strawberry/Orion" Lineage:** After the iterative updates of late 2025, a 5.5 release would likely represent the full integration of more advanced reasoning loops and perhaps the "Dioptra-style" flexible runtimes people have been speculating about.
### What to Watch For
If GPT-5.5 is indeed imminent, the technical community is going to be looking for:
 * **Active Inference:** If it can "think" before it speaks more efficiently than the current O-series.
 * **System 2 Thinking:** True architectural improvements in how it handles complex, multi-step logic.
 * **Reduced Latency:** Whether they’ve managed to scale the reasoning without the 10-30 second wait times.
It’s definitely a "popcorn" week for anyone following model architectures. Are you leaning toward the April 23rd bet, or do you think the 8.3% "No release" crowd might be onto something?
1
7
55
Julián Posada
@posada.website
about 5 hours ago
#PlatformExtractivism is available for 𝐩𝐫𝐞𝐨𝐫𝐝𝐞𝐫 (out Oct 20)! Thrilled to share what @geoplace.bsky.social has to say about the book: "A grounded and uncompromising account of the infrastructures and workers that sustain AI." 𝐋𝐢𝐧𝐤: www.ucpress.edu/books/pla…. #BookRecommendation #Books
This image features a vibrant yellow promotional graphic for a book titled "Platform Extractivism: Data Work and the People Powering Artificial Intelligence" by Julián Posada.

The top of the graphic displays the text "PRE-ORDER NOW! COMING OCTOBER 20, 2026" in bold blue capital letters. Centered in the middle is a 3D rendering of the book cover, which includes three vertical photo strips showing people working in office settings and server racks.

Below the book image is a quote in large, bold black and red font:

"A GROUNDED AND UNCOMPROMISING ACCOUNT OF THE INFRASTRUCTURES AND WORKERS THAT SUSTAIN ARTIFICIAL INTELLIGENCE."
— Mark Graham, co-author of Feeding the Machine: The Hidden Human Labor Powering A.I.

The bottom of the graphic features the logo and name for the University of California Press.
1
6
22
David Bau
@davidbau.bsky.social
about 5 hours ago
2026 is a whirlwind year for AI. Underlying it all is the greatest scientific mystery of our age. How does a neural network think? I talked w Oliver Whang in NYTimes Magazine, on how AI interpretability is a tangle of structure waiting to be unraveled: www.nytimes.com/2026/04/15/...
2
1
8
Digital Humanities @ King's College London
@kingsdh.bsky.social
about 6 hours ago
What happens when AI searches the web for you? Our own Janna Joceli Omena @jannajoceliomena.bsky.social examines GPT-4o's web search over time — and finds that citation diversity only appears when users know how to push the system. 👉 blogs.lse.ac.uk/impactofs…. via @lseimpactblog.bsky.social
Before AI agents act for us, we need to know how AI searches for us - LSE Impact

blogs.lse.ac.uk

Before AI agents act for us, we need to know how AI searches for us - LSE Impact

Generative AI agents are pitched as being a new gateway to engaging with the Internet, but the way AI sees the Internet is both "conservative" and "stubborn”.

0
5
14
Shashank Joshi
@shashj.bsky.social
about 10 hours ago
I wrote something last week on UK nuclear dependence on the US. Experimenting with Claude to see how it handles the same subject. Asked it to produce an interactive tool to catalogue these dependencies. Not a bad effort. Others can poke around it here. splendorous-scone-b67434.netlify.app
1
5
39
rev. howard arson
@theophite.bsky.social
about 13 hours ago
machine learning papers assume a lot of things are measure zero for a field which exclusively works in 32-bit floating point numbers, where measure-zero things have an actual width
4
0
60
Thorne 🌸
@ens0.me
about 14 hours ago
it's good to ask the AI to audit a codebase carefully fairly often, I find it always tends to find actionable items if you haven't done it in a bit, even if some are a bit nitpicky
1
1
24
Ethan Mollick
@emollick.bsky.social
about 16 hours ago
The imaginary optimal selfish scenario for OpenAI, in retrospect, was to keep Reasoners a secret, skip releasing o1 and o1-preview, and release o3 as GPT-5 There would have been no Deep Seek moment, other labs may not have discovered Reasoners quickly, and OpenAI's lead would have been hard to beat
7
0
35
Aaron Sterling
@aaronsterling.bsky.social
about 16 hours ago
I think there's almost a 1:1 skill transfer from mentoring people to skilled LLM use. If you can supervise and train someone skilled & confident but with big blind spots, you can use an LLM well. Patience, clear instructions, expecting things won't be perfect on the first try, build skill over time.

im not exactly convinced you can reskill into AI. what does that even mean? the "skills issues" points to things like curiosity, epistemology, domain expertise, good judgement, taste, etc which were all things we largely failed to systematically skill over the last two decades

0
3
43
Sung Kim
@sungkim.bsky.social
about 17 hours ago
A team at University College London wrote a paper on leaked Claude Code code. Paper: arxiv.org/abs/2604.14228 Repo: github.com/VILA-Lab/Div...
3
5
23
Simon Willison
@simonwillison.net
about 17 hours ago
I upgraded my Claude token counter tool to compare different models and Opus 4.7 appears to use 1.46x times the tokens for text and up to 3x the tokens for images - it's priced the same as Opus 4.6 on a per-token basis so this is actually a pretty big price bump simonwillison.net/2026/Ap….
Screenshot of a token comparison tool with an uploaded screenshot PNG image. Models to compare: claude-opus-4-7 (checked), claude-opus-4-6 (checked), claude-opus-4-5, claude-sonnet-4-6, claude-haiku-4-5. Note: "These models share the same tokenizer". Blue "Count Tokens" button. Results table — Model | Tokens | vs. lowest. claude-opus-4-7: 4,744 tokens, 3.01x (yellow badge). claude-opus-4-6: 1,578 tokens, 1.00x (green badge).
11
19
151
AI Liker Georg
@avengingfemme.bsky.social
about 18 hours ago
i have finally hooked up Claude to a Bluesky MCP thanks to the serendipity of running into @lizthegrey.com's post about the one she set up to do OAuth i feel like i just got bitten by a radioactive spider and incredible new powers are flowing throw me
2
1
25
Vincent Carchidi
@vcarchidi.bsky.social
about 23 hours ago
I don't live by this myself, so there's that, but useful to remind ourselves: whether an LLM/whatever can displace workers is a managerial question, not a technical question.
2
4
17
Thibault Prévost
@winstonsmith.eurosky.social
about 23 hours ago
"Altman ran the venture firm Y Combinator before becoming OpenAI CEO and used that role to build a personal investment portfolio totaling hundreds of startups, a scale rivaling large venture firms. Some of those companies have since struck lucrative deals with OpenAI, enriching Altman."
Sam Altman’s Side Hustles Blur the Line Between OpenAI’s Interests and His Own

www.wsj.com

Sam Altman’s Side Hustles Blur the Line Between OpenAI’s Interests and His Own

Ahead of a planned IPO, Altman’s personal investments remain opaque, making it hard to spot any conflicts.

2
6
4
Ethan Mollick
@emollick.bsky.social
1 day ago
The continuing gap between the capabilities of Gemini Pro 3.1 (very good model) and the capabilities of the Gemini app/website is odd. The model can do what Claude/GPT can do, but there is a minimal harness for tools (file creation, research etc), no auditable CoT/actions, manual canvas, etc. 1/
7
4
67