Activity
Mon
Wed
Fri
Sun
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
What is this?
Less
More

Memberships

JUSTANOTHERPM

983 members • Free

7 contributions to JUSTANOTHERPM
AIPMA | Module 1 Activity | Coh 001
Please share a document with the LLM's name, prompt and the learning summary of session. Please include a visual (optional) Also share in the comments below how would you define "good quality" in this case, and how would you measure success of the "Online classes learning summariser" feature
1 like • 4d
Good Quality Control (What does “good quality” actually mean for this summarizer?) 1. Source Fidelity (Truth Over Fluency). Every output must be traceable back to the transcript. No invented concepts, no “helpful” additions, no synthesized frameworks. Fluency is secondary to faithfulness. If the speaker didn’t say it, it doesn’t exist. Quality starts with strict grounding. 2. Concept Extraction (Frameworks Beat Filler). The summarizer must prioritize mental models, decisions, and product insights over conversational noise. Logistics and small talk should disappear. Core frameworks (like deterministic vs. probabilistic) must surface clearly. If filler survives but concepts don’t, extraction failed. 3. Applied Insight (From Notes to Leverage). A strong summary converts ideas into practical implications. It should answer: What changes for me as a PM? If users finish reading without clearer direction on discovery, design, or measurement, the summarizer produced notes, not value. 4. Cognitive Efficiency (Designed for Two-Minute Recall). Formatting is part of quality. Use structured sections, crisp bullets, and scannable layouts so users can refresh the entire session in under two minutes. The goal isn’t completeness—it’s fast comprehension. Measuring Success: 1. Friction Index (How Much Did Users Have to Fix?). Measure how much users modify the output before keeping or sharing it. Minimal edits mean the summary matched intent. Heavy rewrites or deletions signal quality gaps. Low friction = real time saved. 2. Reuse Signal (Did It Become Working Material?). Track whether users copy sections into docs, notes, or follow-ups. When content leaves the product and shows up in real workflows, that’s stronger than any rating—it proves usefulness. 3. Steerability Rate (Can Users Recover Quickly?). Measure how often users successfully improve a weak result using regenerate, focus, or refinement controls. If users can course-correct and accept the next output, your recovery UX is doing its job. 4. Reference Benchmarking (Are We Matching Expert Output?). Maintain a small set of expert-written “reference summaries.” Regularly score AI outputs against them on coverage, accuracy, and actionability. This gives you a concrete baseline to track real improvement over time.
Week 2 Activity 1: What tech stack does your product need
Submit your answer here. Keep it simple. Just explain in simple English. Be sure to call out "why" you think you need or don't need a specific aspect in your product. Let's go 👇
1 like • 21d
@Akshun Gulati This is a great problem to solve. I see many places where a solution like this would be impactful. Bad invoices and purchase orders information can lead to expensive problems for companies. One of the big blockers you would need to overcome is the level of accuracy of the results, (e.g., many companies have penalties associated with missing due dates so you won't want to mis-read a due date).
1 like • 16d
@Jerel Lee, I think you can do it without RAG, but I see your point. If you can get access to the APIs for the Applicant Tracking Systems, like Workday, etc. you can verify sources and accurate listings with an API call. But I can see the value of RAG for keeping track of this information.
Week 4 Activity
Look at a product of your choice and apply the AI PM lens to it.
2 likes • 20d
The Framework for AI PM lens. Product of choice: ChatGPT Use Case: summarizing long reports and documents, e.g. NPS report, with ChatGPT Use these six questions to guide your thinking: 1. What's the real job this solves? Not what the AI does technically. What does the user actually need? Ans: The user needs to save time in extracting the key information out of documents so that they can share with stakeholders or look good in front of their management team. 2. How does the system stay grounded? What keeps it from making stuff up? (Data? Limited scope? Sources? Guardrails?) Ans: The system uses the data that the user provides. In this case it would be documents or reports to summarize. It can also show reasoning if asked by the user. 3. What's the context the model receives? What instructions or information did the team give the model about how to behave? Ans: The instructions were clear in the user prompt, “summarize this NPS report and provide the top five most critical users concerns and recommended follow-up action items.” If the output from the system did not provide clear info, then the user would follow-up with a refined prompt. 4. What are the failure modes? What could go wrong? And when it does, how does the product handle it? Ans: The user could provide a vague prompt, like summarize this report. In this case, the system might not give the user the specific summary that they are looking for. The other thing that could go wrong is the user ask for specific information from the documents, but the AI system could not find that information. In this case the AI system will either hallucinate or say “sorry I don’t have that information.” 5. What trade-offs did the team make? Did they choose speed or accuracy? Consistency or flexibility? Safety or usefulness? How can you tell? Ans: They made the following trade-offs: -- They picked speed over accuracy because the system answers instantly -- They picked flexibility over consistency because the system sometimes give different answers
Week 3 Activity: Does it really need AI
For the idea that you thought of in the Week 2 activity, share the following: Deliverable #1: Share the scores on each dimension and share a short description of why you rated it like that. Deliverable #2: Share the total score (Total Score = add all three) Deliverable #3 : What does your score tell you about your idea?
0 likes • 21d
The Three Dimensions Of AI Fit Score on the AI job finder for PMs who are in transition idea. Data Readiness: Score = 4: Good data, but some gaps. Lot of Job listings, and resumes, but you have to find a good way to aggregate them because you can’t use the job boards APIs. Output type: Score = 5: Subjective/judgment-based (AI is built for this.). The output is a conversation on what’s right for an individual based on their experience, skill-sets, goals and a set of job listings. Error tolerance: Score = 3: Errors acceptable with human review or correction (Human-in-the-loop works.). If the LLM gets the recommendation wrong, the user can correct refine it and ask a better prompt or provide more context. The user will be tolerant of errors because they view this as exploratory. Overall Score = 12: This is a strong AI candidate based on the overall fit score falling between 11 and 15. AI is a great solution for this type of user exploratory problem where users want to talk with the AI to get better ideas on how to position themselves for certain job roles.
0 likes • 21d
@Peculiar Ediomo-Abasi I see data readiness and error tolerance as your two biggest challenges. I'm guessing error tolerance would be the biggest blocker because errors could be expensive for both hospitals and patients. This is going to be an interesting area to explore with hypothesis testing, e.g., will stakeholders handle the potential errors for upside benefits or do they fall back to old solutions?
Week 1, Activity 2: Personal Inventory
Submit your problem mapping here. 👇 How to Submit 1. Fill out the template from the essay 2. Post your response in the comments below 3. Read at least 2 other people's ideas and leave thoughtful feedback. Let's think this through. 👇
0 likes • 23d
@Masahiro Teramoto, this is an interesting problem to solve. I often ask myself whether my learnings are translating into meaningful growth. It's hard to see sometimes especially if the improvement is gradual over time. A system that provides a set of reference criterias and feedback over time would help me to have someway of judging my progress. One of the challenges is the system would need to find creative ways to help me stay encouraged, e.g., think Duolingo versus traditional language learning programs.
1 like • 23d
@Peculiar Ediomo-Abasi I see a similar problem in a number of countries. AI has made good progress in predictive analytics within big industries and companies. Largely due to those places having lots of good data. The challenge I see we would need to overcome in health care is getting the data and ensuring continuous discipline in feedback loops(more good data) into the system. It would also take time to get this system to be good enough. People would need to have the right expectations about what the system can do in the short versus mid term.
1-7 of 7
Phil L
2
6points to level up
@phil-l-6559
Product Management leader in real estate technology.

Active 4d ago
Joined Jan 11, 2026
Powered by