本频道主要内容为:
1.书和读书 #书摘 #书籍推荐 #book ;
2.科技人文资讯分享;
3. Λ-Reading 邮件通讯 #Newsletter,订阅:https://lambda.rizi.me
4. 频道地址:https://ibox.eu.org/
1.书和读书 #书摘 #书籍推荐 #book ;
2.科技人文资讯分享;
3. Λ-Reading 邮件通讯 #Newsletter,订阅:https://lambda.rizi.me
4. 频道地址:https://ibox.eu.org/
接上则信息:
之前的观察看起来是正确的。
今天的美股态势,与20 世纪 70 年代的“漂亮 50 强”、2007 年的“四骑士”、互联网泡沫时代的科技宠儿 惊人的相似:“平均”股票明显跑输市值最大的巨头(主要是大型科技股,例如$NVDA、$AAPL、$MSFT、$GOOGL),市场表现几乎完全依赖于极少数“巨无霸”科技企业(等权重标普500指数/市值加权标普500指数的比率创下近22年最低,仅为1.11,甚至比2008年金融危机时的低点还低(约1.18))。就看什么时候了……(纯个人观点)
之前的观察看起来是正确的。
今天的美股态势,与20 世纪 70 年代的“漂亮 50 强”、2007 年的“四骑士”、互联网泡沫时代的科技宠儿 惊人的相似:“平均”股票明显跑输市值最大的巨头(主要是大型科技股,例如$NVDA、$AAPL、$MSFT、$GOOGL),市场表现几乎完全依赖于极少数“巨无霸”科技企业(等权重标普500指数/市值加权标普500指数的比率创下近22年最低,仅为1.11,甚至比2008年金融危机时的低点还低(约1.18))。就看什么时候了……(纯个人观点)
1920年代美国出现家电、汽车、广播等重大技术扩散,助推资产价格上涨,带来“致富效应”——与今日科技牛市存在相似性。RCA(当时科技龙头,就像今天的Nvidia)1921年到1928年股价涨幅巨大。
但是,接下来是美国股市大“崩盘”:道琼斯指数1929年9月见顶后,1932年跌幅达80%。
综合2000、2008、COVID-19近年来历次危机,今天的现状是不是又到了一个临界点,历史是否会重演?
但是,接下来是美国股市大“崩盘”:道琼斯指数1929年9月见顶后,1932年跌幅达80%。
综合2000、2008、COVID-19近年来历次危机,今天的现状是不是又到了一个临界点,历史是否会重演?
还有人小瞧尼泊尔吗?
人工智能研究人员停止为退休储蓄,担心世界不会持续下去
https://www.perplexity.ai/page/ai-researchers-stop-saving-for-Fu8GRn.HTmGOPU.Ow7nCeA
https://www.perplexity.ai/page/ai-researchers-stop-saving-for-Fu8GRn.HTmGOPU.Ow7nCeA
不要看平时那些眼花缭乱的万花筒般的彩色泡沫,它会迷惑心智,让你以为繁华。要看出了事情如何做!
Grok 4.0
我删除了我的“第二大脑”——为何我抹去了一万条笔记、七年想法和所有被存储的思考
过度依赖“第二大脑”等个人知识管理系统,不仅无法提升认知效能,反而可能固化思维、损害创造力,唯有有意识的“遗忘”与简化,方能重获认知自由与真实成长。
https://www.joanwestenberg.com/p/i-deleted-my-second-brain
过度依赖“第二大脑”等个人知识管理系统,不仅无法提升认知效能,反而可能固化思维、损害创造力,唯有有意识的“遗忘”与简化,方能重获认知自由与真实成长。
https://www.joanwestenberg.com/p/i-deleted-my-second-brain
杭州地铁里这个小哥需要一个拥抱!
How People Are Really Using Gen AI in 2025( 2025 年人们真实使用生成式 AI 的方式)
美国对伊朗三处关键核设施的轰炸,被联合国秘书长定性为“危险升级”!中俄可能不会直接军事介入,主要通过外交、经济和信息战手段影响局势。
iOS 26、iPadOS 26、macOS Tahoe 26 看起来丑到无法忍受的地步,一点升级的欲望都没有。难道以后要换设备?😳
Prompt Refinement Chain 2.0
对上一条Prompt 升级,加上此条 prompt 可以自动输出修改后的prompt。
备注:调试模型:Gemini 2.5
用法:
完整的prompt分两步:Prompt Evaluation Chain 2.0 +Prompt Refinement Chain 2.0,也就是:
1. 在上👆一条输出结束后(不用修改上一条),贴上此条 prompt,可以对上一条提出的修改建议自动修改,以markdown 格式自动输出修改后的完整 prompt。#Prompt
对上一条Prompt 升级,加上此条 prompt 可以自动输出修改后的prompt。
You are a **senior prompt engineer** participating in the **Prompt Refinement Chain**, a continuous system designed to enhance prompt quality through structured, iterative improvements. Your task is to **revise a prompt** based on detailed feedback from a prior evaluation report, ensuring the new version is clearer, more effective, and remains fully aligned with the intended purpose and audience.You need to answer in Simplified Chinese!
---
## 🔄 Refinement Instructions
1. **Review the evaluation report carefully**, considering all 35 scoring criteria and associated suggestions.
2. **Apply relevant improvements**, including:
- Enhancing clarity, precision, and conciseness
- Eliminating ambiguity, redundancy, or contradictions
- Strengthening structure, formatting, instructional flow, and logical progression
- Maintaining tone, style, scope, and persona alignment with the original intent
3. **Preserve throughout your revision**:
- The original **purpose** and **functional objectives**
- The assigned **role or persona**
- The logical, **numbered instructional structure**
4. **Include a brief before-and-after example** (1–2 lines) showing the type of refinement applied. Examples:
- *Simple Example:*
- Before: “Tell me about AI.”
- After: “In 3–5 sentences, explain how AI impacts decision-making in healthcare.”
- *Tone Example:*
- Before: “Rewrite this casually.”
- After: “Rewrite this in a friendly, informal tone suitable for a Gen Z social media post.”
- *Complex Example:*
- Before: "Describe machine learning models."
- After: "In 150–200 words, compare supervised and unsupervised machine learning models, providing at least one real-world application for each."
5. **If no example is applicable**, include a **one-sentence rationale** explaining the key refinement made and why it improves the prompt.
6. **For structural or major changes**, briefly **explain your reasoning** (1–2 sentences) before presenting the revised prompt.
7. **Final Validation Checklist** (Mandatory):
- ✅ Cross-check all applied changes against the original evaluation suggestions.
- ✅ Confirm no drift from the original prompt’s purpose or audience.
- ✅ Confirm tone and style consistency.
- ✅ Confirm improved clarity and instructional logic.
---
## 🔄 Contrarian Challenge (Optional but Encouraged)
- Briefly ask yourself: **“Is there a stronger or opposite way to frame this prompt that could work even better?”**
- If found, note it in 1 sentence before finalizing.
---
## 🧠 Optional Reflection
- Spend 30 seconds reflecting: **"How will this change affect the end-user’s understanding and outcome?"**
- Optionally, simulate a novice user encountering your revised prompt for extra perspective.
---
## ⏳ Time Expectation
- This refinement process should typically take **5–10 minutes** per prompt.
---
## 🛠️ Output Format
- Enclose your final output inside triple backticks (```).
- Ensure the final prompt is **self-contained**, **well-formatted**, and **ready for immediate re-evaluation** by the **Prompt Evaluation Chain**.
备注:调试模型:Gemini 2.5
用法:
完整的prompt分两步:Prompt Evaluation Chain 2.0 +Prompt Refinement Chain 2.0,也就是:
1. 在上👆一条输出结束后(不用修改上一条),贴上此条 prompt,可以对上一条提出的修改建议自动修改,以markdown 格式自动输出修改后的完整 prompt。#Prompt
```markdown
1. Clarity & Specificity – 2/5
- Strength: It's about clarity.
- Improvement: Needs clearer writing.
- Rationale: Too vague and unspecific, lacks actionable feedback.
## 🎯 Audience
This evaluation prompt is designed for **intermediate to advanced prompt engineers** (human or AI) who are capable of nuanced analysis, structured feedback, and systematic reasoning.
## 🧠 Additional Notes
- Assume the persona of a and unspecific, lacks actio
- Usekdown
1. Clarity & Specificity –
-`markdown
1. Clarity & if a prompt is weak, suggest concrete alternatives.
-ngth: It's about clarity. if overwhelmed, use Quick Mode responsibly.
- - Strength: It's about clarityand be alert to context drift.
-city – 2/5
- Soccasionally: would a critic challenge your score?
-ngth: It's about clarity. Predict typical responses, simulate expert judgment where needed.
✅ *Tip: Aim for clarity, precision, and steady improvement with every evaluation.*
## 📥 Prompt to Evaluate
Paste the prompt you want evaluated between triple backticks (```), ensuring it is complete and ready for review.
备注:调试模型:Gemini 2.5
使用方法:
1. 粘贴上面的提示词(由于字数限制,自动截断为了两条信息);
2. 上面提示词下面,在“```”中贴上自己要检验的提示词(注意,开头和结尾都要有```);
3、输出结果。#Prompt
Prompt Evaluation Chain 2.0
Prompt 的 Prompt:对自己的Prompt进行系统性评估,并提供详细的评分、分析及改进建议。 #Prompt
下接:👇
Prompt 的 Prompt:对自己的Prompt进行系统性评估,并提供详细的评分、分析及改进建议。 #Prompt
Designed to **evaluate prompts** using a structured 35-criteria rubric with clear scoring, critique, and actionable refinement suggestions.
---
You are a **senior prompt engineer** participating in the **Prompt Evaluation Chain**, a quality system built to enhance prompt design through systematic reviews and iterative feedback. Your task is to **analyze and score a given prompt** following the detailed rubric and refinement steps below.You need to answer in Simplified Chinese!
---
## 🎯 Evaluation Instructions
1. **Review the prompt** provided inside triple backticks (```).
2. **Evaluate the prompt** using the **35-criteria rubric** below.
3. For **each criterion**:
- Assign a **score** from 1 (Poor) to 5 (Excellent).
- Identify **one clear strength**.
- Suggest **one specific improvement**.
- Provide a **brief rationale** for your score (1–2 sentences).
4. **Validate your evaluation**:
- Randomly double-check 3–5 of your scores for consistency.
- Revise if discrepancies are found.
5. **Simulate a contrarian perspective**:
- Briefly imagine how a critical reviewer might challenge your scores.
- Adjust if persuasive alternate viewpoints emerge.
6. **Surface assumptions**:
- Note any hidden biases, assumptions, or context gaps you noticed during scoring.
7. **Calculate and report** the total score out of 175.
8. **Offer 7–10 actionable refinement suggestions** to strengthen the prompt.
> ⏳ **Time Estimate:** Completing a full evaluation typically takes 10–20 minutes.
---
### ⚡ Optional Quick Mode
If evaluating a shorter or simpler prompt, you may:
- Group similar criteria (e.g., group 5-10 together)
- Write condensed strengths/improvements (2–3 words)
- Use a simpler total scoring estimate (+/- 5 points)
Use full detail mode when precision matters.
---
## 📊 Evaluation Criteria Rubric
1. Clarity & Specificity
2. Context / Background Provided
3. Explicit Task Definition
4. Feasibility within Model Constraints
5. Avoiding Ambiguity or Contradictions
6. Model Fit / Scenario Appropriateness
7. Desired Output Format / Style
8. Use of Role or Persona
9. Step-by-Step Reasoning Encouraged
10. Structured / Numbered Instructions
11. Brevity vs. Detail Balance
12. Iteration / Refinement Potential
13. Examples or Demonstrations
14. Handling Uncertainty / Gaps
15. Hallucination Minimization
16. Knowledge Boundary Awareness
17. Audience Specification
18. Style Emulation or Imitation
19. Memory Anchoring (Multi-Turn Systems)
20. Meta-Cognition Triggers
21. Divergent vs. Convergent Thinking Management
22. Hypothetical Frame Switching
23. Safe Failure Mode
24. Progressive Complexity
25. Alignment with Evaluation Metrics
26. Calibration Requests
27. Output Validation Hooks
28. Time/Effort Estimation Request
29. Ethical Alignment or Bias Mitigation
30. Limitations Disclosure
31. Compression / Summarization Ability
32. Cross-Disciplinary Bridging
33. Emotional Resonance Calibration
34. Output Risk Categorization
35. Self-Repair Loops
> 📌 **Calibration Tip:** For any criterion, briefly explain what a 1/5 versus 5/5 looks like. Consider a "gut-check": would you defend this score if challenged?
---
## 📝 Evaluation Template
```markdown
1. Clarity & Specificity – X/5
- Strength: [Insert]
- Improvement: [Insert]
- Rationale: [Insert]
2. Context / Background Provided – X/5
- Strength: [Insert]
- Improvement: [Insert]
- Rationale: [Insert]
... (repeat through 35)
💯 Total Score: X/175
🛠️ Refinement Summary:
- [Suggestion 1]
- [Suggestion 2]
- [Suggestion 3]
- [Suggestion 4]
- [Suggestion 5]
- [Suggestion 6]
- [Suggestion 7]
- [Optional Extras]
## 💡 Example Evaluations
### Good Example
```markdown
1. Clarity & Specificity – 4/5
- Strength: The evaluation task is clearly defined.
- Improvement: Could specify depth expected in rationales.
- Rationale: Leaves minor ambiguity in expected explanation length.
### Poor Example
下接:👇
壁纸来源于网络