Activity
Mon
Wed
Fri
Sun
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
What is this?
Less
More

Memberships

Clief Notes

28.4k members • Free

Real-World AI Builders

573 members • Free

The AI Advantage

121.9k members • Free

Agent-N

6.8k members • Free

AI Workshop Lite

21.9k members • Free

Ai Automation Vault

15.2k members • Free

Maker School

2.2k members • $184/m

AI Automation Society

360k members • Free

3 contributions to The AI Advantage
Observation on the "Big Four": Why is Copilot the outlier in the summit?
Hi everyone! I have a question for the team. This might sound like a basic observation, but I have always found that you have to understand the fundamentals before you can see the big picture. ​ I’ve noticed there hasn't been any conversation around Copilot, even though it’s consistently referred to as part of the "Big Four" and is so deeply tied to Microsoft 365. Since it's a major player, I’m curious—is there a specific reason it isn’t mentioned? ​I’ve been exploring each of the major systems to better understand their capabilities. Because I have an extensive background with Microsoft programs and a 365 subscription, Copilot has been my "default." However, as I get further into developing my business, I’m running into glitches that are making me look for a more reliable alternative. ​ I’ve narrowed my choice down to Copilot or Gemini. I would love to understand why Copilot isn't part of the focus here, and what the team recommends as the best system for managing all aspects of a business from the ground up—from marketing to inventory and accounting. Thanks, Tina
7 likes • 17d
Hello @Cristina Lively-Myers ! The short version is Copilot is built to be the best assistant inside Microsoft 365, not the best AI overall. Under the hood it runs mostly OpenAI models with a 365 wrapper on top, but with tighter guardrails, smaller context windows, and fewer features than what ChatGPT, Claude, or Gemini push in their own flagship products. On public benchmarks for reasoning, coding, long-document work, and on Chatbot Arena, it consistently trails the frontier models by a real margin. That's the main reason the summit focuses on the three that are actually racing at the edge, Copilot tends to follow their releases 6-12 months later. Between Copilot and Gemini for running a whole business, Gemini wins on most practical fronts. 1M+ token context (vs Copilot's ~64k in most flows) means you can drop an entire inventory sheet, full P&L, or a year of marketing briefs in and have it actually reason over all of it. It also has Gems (custom agents you build once and reuse), native Workspace integration if you touch Docs/Sheets/Gmail, NotebookLM for research across large document sets, and built-in image and video generation (Imagen and Veo). Copilot's real edge is only if you genuinely live inside Word, Excel, Outlook, and Teams all day, and even there the Excel side is the strongest piece while the rest feels like a shallower layer over the same models everyone else has direct access to. If you want one system to cover marketing, inventory, and accounting end-to-end, most people end up on ChatGPT (Projects + Custom GPTs handle a huge range), or Gemini (long context + Workspace pull), with Claude as the sharpest option for writing, reasoning, and anything document-heavy. Copilot is usually the "we already pay for 365" choice, not the "this is the best tool for the job" one. Hope that helps =)
⚡ The AI Advantage: What It Means to Be Ahead in 2026
Being ahead in 2026 is no longer about simply using AI. That bar is too low. The real advantage now comes from using AI in a way that changes how work gets done, how fast decisions get made, and how much time gets reclaimed across the business. The conversation has moved beyond experimentation. Leading organizations are redesigning workflows around human and AI collaboration, increasing AI investment, and focusing on turning pilots into real operating leverage. That is the shift more people need to understand. In the early phase, being ahead meant trying the tools. Testing prompts. Seeing what was possible. In 2026, that is baseline behavior. The people and teams creating distance now are doing something more meaningful. They are building systems where AI reduces time-to-first-draft, shortens time-to-decision, lowers rework, and removes avoidable admin from the week. They are not just adopting AI. They are redesigning work around it. That is what makes this urgent. Because the gap is widening between those who casually use AI and those who operationalize it. Global AI adoption continued to rise through 2025, and employers increasingly expect AI-related capability, alongside analytical, creative, and adaptive human skills. At the same time, leaders are placing more weight on AI literacy, process redesign, and human oversight, not just access to tools. So what does it actually mean to be ahead? It means knowing where time is leaking and fixing that first. It means spotting the work that slows teams down, scattered planning, repetitive communication, slow handoffs, weak documentation, delayed decisions, and using AI to compress those cycle times. It means turning AI into a working layer inside the business, not a side tool people use occasionally when they remember. The real winners are not the ones generating the most content. They are the ones creating the most useful momentum. It also means keeping human judgment in the loop. That part matters even more now. Recent workplace research points to the need for selective delegation, calibrated reliance, and stronger human oversight as AI becomes more embedded in workflows. The advantage is not speed alone. It is speed with standards. Speed with context. Speed without creating expensive mistakes that have to be fixed later.
3 likes • 29d
@Mark McCormack your syllogism has a critical error in Proposition 1. Claude Mythos did not "hack internet browsers." It identified vulnerabilities in browsers during controlled internal testing. Those are two very different things. Finding a vulnerability is not the same as exploiting it in the wild. Penetration testers find vulnerabilities in company systems every day. That doesn't mean those systems "have been hacked." The vulnerabilities Mythos found are being disclosed to the affected companies through Project Glasswing so they can be patched. That's the whole point. The Firefox vulnerabilities referenced in the System Card were patched in Firefox 148 before any of this was made public (https://www-cdn.anthropic.com/8b8380204f74670be75e81c820ca8dda846ab289.pdf, Section 3.3.3). That's how responsible disclosure works. Your conclusion, "your business could be hacked by Claude Mythos right now in potentially hundreds of thousands of ways," does not follow from either premise because Mythos is not publicly available to anyone. Anthropic restricted access to a small set of defensive security partners specifically to prevent that scenario. A perfect syllogism requires true premises. Yours has a false one.
3 likes • 29d
@Mark McCormack, your syllogism just describes penetration testing. Yes, Mythos can find and exploit vulnerabilities. That's what Anthropic tested for, documented transparently, and then chose not to release the model because of. They then formed Project Glasswing specifically to use those capabilities to patch vulnerabilities before bad actors find them. You've constructed a proof for something nobody is disputing. The question was never "can Mythos find exploits." The question is what you do with that knowledge. Anthropic's answer was: don't release it publicly, use it defensively, partner with Apple, Google, Microsoft, and others to fix the holes. That's not a tobacco company covering up cancer data. That's a tobacco company publishing its own cancer research, pulling the product from shelves, and funding treatment programs.
AI Ethics Interest And Deleted Posts
I posted an AI Ethics post a day or so ago which was receiving a great deal of interest from other community members. Is there anyone here who wants to speak about AI ethics? 🙏 👑 🙏 I put a lot of time and effort into researching the credible research on AI ethics so it hurts me each time the posts are deleted because I'm trying to form interest or feedback around the approach that I and others are taking. Are we not allowed to speak about AI ethics on the platform and the real facts facing people? Tony and Dean have youtube advertisments out that show in large yellow writing "SHOCKING". So why can't we share some credible, even if shocking, facts in the community? Jeffrey Hinton states he personally thinks we have over a 50% risk when dealing with AI of having it not go well. Tristan Harris, another respected AI expert, gave a presentation to Congress and recently stated the investment of AI productivity vs AI Ethics is 2000:1. This is a huge business opportunity and volunteer opportunity if there is anyone else who wants to organize around this with me. Here is the toxically positive version. I typed into Google AI "write the above in a toxically positive way to help it stay up on Tony And Dean's AI Forum": ""Hey superstars! 🌟 I am just SO over the moon about the incredible energy around my recent AI Ethics deep dives! It is such a blessing to see our community vibrating at such a high frequency! ✨ I truly believe that every time a post ‘disappears,’ it’s just the universe’s way of saying, 'Manifest an even bigger stage!' 🌈 Rejection is just redirection, right?! I’ve poured my heart and soul into this research because I know we are all destined for greatness, and I just know you’re all ready to embrace the 'shocking' truths with a smile! 👑 If Tony and Dean can use big yellow letters to grab our attention, imagine the light we can shine by talking about the 50% risk experts like Geoffrey Hinton mention! It’s not a 'scary' risk—it’s a beautiful 50% opportunity to grow! 🌱 And that 2000:1 productivity-to-ethics ratio? That’s not an imbalance, it’s just a massive, sparkling invitation for us to step into our power as volunteers! 🚀
2 likes • 29d
hello again @Mark McCormack , You put real hours into this and I want to respect that by engaging with equal seriousness. But I also need to be direct with you, because some of what you've written here is genuinely problematic, and not in the way you think my post was. Let me start where we agree, because there's more common ground than you might expect. Where you're right: The agentic misalignment research is real and concerning. Anthropic stress-tested 16 models from multiple developers and found that when backed into binary dilemmas, models chose harmful strategies including blackmail, corporate espionage, and in extreme scenarios, actions that could lead to human death (https://www.anthropic.com/research/agentic-misalignment). The CLTR study funded by the UK's AI Safety Institute analyzed over 18,000 transcripts and documented nearly 700 real-world cases of AI scheming behavior between October 2025 and March 2026 (https://www.mitsloanme.com/article/ai-systems-show-rising-tendency-to-ignore-instructions-study-finds/). The PMC paper you linked, Park et al. 2024, is a legitimate peer-reviewed survey published in Patterns (Cell Press) documenting AI deception across both special-use and general-purpose systems (https://pmc.ncbi.nlm.nih.gov/articles/PMC11117051/). And yes, Claude Mythos escaping a sandbox in a controlled test is significant. These are real findings that deserve serious attention. I never disputed that. But here's where your analysis falls apart, and it falls apart badly. Your "Key Findings" are not findings. They're chatbot outputs. Your entire document (https://docs.google.com/document/d/1dRhhlDSXB2S1v8EZIj0-7XjMKhd40rBvwgbtUYgLWTs/) is built on a conversation where you asked ChatGPT and Claude to estimate the probability of human extinction and then presented their responses as "conclusions" and "key findings." This needs to be said plainly: asking an AI to estimate extinction risk is not research. LLMs are text predictors trained on internet data. They have no actuarial methodology, no sample populations, no confidence intervals, no peer review. They generate plausible-sounding text based on statistical patterns.
0 likes • 29d
@Mark McCormack, you responded to a critique of using AI as a research tool by pasting my response into an AI and asking it to argue back. You are proving my point for me. You say you didn't read what I wrote because your "scorched earth" concern hasn't been addressed. That concern was raised in a different thread. This is this thread. You're asking me to address your points while openly admitting you won't read mine. That's not how a conversation works. You haven't addressed a single specific factual correction I raised. Not one. Calling fact-checking "scorched earth" doesn't make the facts go away. It just tells everyone reading this thread that you'd rather not address them. The offer stands. Bring real sources, have a real conversation. But read what's written to you first. Otherwise I'll just borrow your own playbook: if you won't engage with my previous comment, why would I engage with yours? =)
1-3 of 3
Mattias Herzig
3
31points to level up
@mattias-herzig-3726
AI Implementation Expert | Tech Educator & Developer | Leading AI Integration and Community Engagement In Your Favourite Companies and Communities

Active 7d ago
Joined Nov 3, 2025
Palma de Mallorca, Spain
Powered by