公式動画ピックアップ

AAPL   ADBE   ADSK   AIG   AMGN   AMZN   BABA   BAC   BL   BOX   C   CHGG   CLDR   COKE   COUP   CRM   CROX   DDOG   DELL   DIS   DOCU   DOMO   ESTC   F   FIVN   GILD   GRUB   GS   GSK   H   HD   HON   HPE   HSBC   IBM   INST   INTC   INTU   IRBT   JCOM   JNJ   JPM   LLY   LMT   M   MA   MCD   MDB   MGM   MMM   MSFT   MSI   NCR   NEM   NEWR   NFLX   NKE   NOW   NTNX   NVDA   NYT   OKTA   ORCL   PD   PG   PLAN   PS   RHT   RNG   SAP   SBUX   SHOP   SMAR   SPLK   SQ   TDOC   TEAM   TSLA   TWOU   TWTR   TXN   UA   UAL   UL   UTX   V   VEEV   VZ   WDAY   WFC   WK   WMT   WORK   YELP   ZEN   ZM   ZS   ZUO  

  公式動画&関連する動画 [Pop Goes the Stack | LLM-as-a-Judge: Bias, Preference Leakage, and Reliability | AI Bias]

We're back with another episode of Pop Goes the Stack and the newest bright idea in #AI: don’t pay humans to evaluate model outputs, let another model do it. This is the “LLM-as-a-judge” craze. Models not just spitting answers but grading them too, like a student slipping themselves the answer key. It sounds efficient, until you realize you’ve built the academic equivalent of letting someone’s cousin sit on their jury. The problem is called preference leakage. Li et al. nailed it in their paper “Preference Leakage: A Contamination Problem in LLM-as-a-Judge.” They found that when a model judges an output that looks like itself—same architecture, same training lineage, or same family—it tends to give a higher score. Not because the output is objectively better, but because it “feels familiar.” That’s not evaluation, that’s model nepotism. Watch as #F5's Lori MacVittie, Joel Moses, and Ken Arora explore the concept of preference leakage in AI judgement systems. Tune in to understand the risks, the impact on the enterprise, and actionable strategies to improve model fairness, security, and reliability. Chapters: 00:00 Welcome to Pop Goes the Stack 00:34 LLM-as-a-judge and preference leakage 01:26 Why are judgement systems necessary? 03:05 Bias in judgment systems and model families 06:56 Is AI bias a problem or a feature? 08:13 Preference leakage vs data leakage 10:00 Impact of synthetic data: Generation and model training 12:07 Evaluating models: Red teaming and diversity 14:44 What can we do about preference leakage? 17:21 SLMs, preference leakage, and protecting data transactions 19:21 Correctness' implication on the enterprise: GenAI security and reliability 20:45 Key takeaways: Measure carefully and diversify Learn how you can stay ahead of the curve and keep your stack whole with additional insights on app security, multicloud, AI, and emerging tech: https://go.f5.net/i4g89z0k Read the paper, Preference Leakage: A contamination Problem in LLM-as-a-judge: https://go.f5.net/eacp2top
 117      3