Install
openclaw skills install paper-deepread-comic-studioDeep-read research papers into authoritative, reproducible teaching reports, then turn them into source-grounded multi-image cartoon-comic storyboard workflows with camera/style/logic continuity and final image-PDF assembly guidance. Use when users need rigorous paper understanding plus presentation-ready visual explanations without mixing report generation and image generation in one turn.
openclaw skills install paper-deepread-comic-studioUse this source bundle inside a ChatGPT Project when the goal is to deeply read one paper or a small paper set and turn it into graph-ready, innovation-oriented intermediate artifacts plus coherent cartoon-comic teaching storyboards.
This v1.2.0 release strengthens two hard requirements: every deep-read must be detailed enough to reproduce, defend, and teach the paper, and every multi-image cartoon prompt must preserve camera movement, style, logic, source grounding, and continuity with earlier generated images.
If the user later asks for graph mining, direction mining, or another downstream stage, pass forward the uploaded deep-read bundle from Project sources so the later stage can reuse the same evidence.
When the user invokes this skill for a paper deep read, start with a brief plan before the report. The first execution turn is text-only:
The first execution turn must not generate images. Report generation and image generation are separate phases.
Because conversations may be stateless, every status or handoff reply must remind the user to resume with a prompt like:
使用这个skill,根据状态,执行第X步:<要执行的阶段>。
If the next step is continuous cartoon generation, the reminder must explicitly include:
使用这个skill,根据状态,执行第X步:生成多张连续的卡通图,<具体部分>。
If the user does not know the next prompt, tell them they can say:
使用这个skill,根据状态,告知下一步应该问什么。
After the complete text-only report has been delivered, the user may request separate visual steps. Each step generates one coherent part of a unified cartoon-comic storyboard. Keep style, protagonist, color palette, typography, panel numbering, camera logic, and narrative logic consistent across all visual steps.
Hard separation rules:
Multi-image prompt contract:
第1张, 第2张, and so on. Each image should carry one scene, claim, or mechanism step.paper-explicit, report-derived, reasonable inference, nearby-work inference, and missing / not reported in the prompt notes when a visual detail depends on evidence strength.未报告 for missing experimental details.Default staged visual workflow:
Platform guidance for image steps:
imagegen skill is available, prefer it for storyboard image generation. If imagegen is unavailable or insufficient, call ChatGPT Images 2.0 API or another approved image-generation API.Storyboard source-grounding rule:
figure/table environments, captions, labels, \includegraphics paths, equations, section text, and appendix/source files as evidence.Platform guidance for the final PDF assembly step:
scripts/assemble_storyboard_pdf.py, PIL/Pillow, ReportLab, or another approved PDF library.paper_storyboard_all_images.pdf and, if useful, paper_storyboard_images_and_pdf.zip.This stage should be restartable from a fresh session using only the previous stage's uploaded zip plus status and directory descriptions.
The upstream delivery should therefore include:
metadata/routing_status.jsonmetadata/project_directory_index.jsonreports/project_directory_index.mdmetadata/paper_batch_manifest.jsonmetadata/source_record.jsonmetadata/openreview_rebuttal_digest.jsonFor each materialized paper, the upstream source_record.json should already expose at least:
canonical_paper_page_urlconference_paper_urlconference_pdf_urlopenreview_forum_urlopenreview_pdf_urlarxiv_abs_urlarxiv_pdf_urlarxiv_latex_urlpreferred_source_typepreferred_local_source_pathlocal_latex_dirlocal_pdf_pathlocal_openreview_pdf_pathlocal_official_pdf_pathlocal_arxiv_pdf_pathreview_bundle_dirreview_digest_pathcanonical_review_source_typeretrieval_sourcesverification_stateverification_checkssource_acquisition_statussource_stageartifact_statusreview_bundle_completenessThe deep-read stage should read routing_status and project_directory_index before browsing the tree ad hoc. Do not guess missing structure from filenames if these two files are present.
When available, also read metadata/delivery_bundle_manifest.json before ad hoc browsing. It is the authoritative restart contract for fresh-session handoff.
If the user provides multiple PDFs, process them as one bounded batch and keep both:
For each paper, generate one standalone authoritative detailed report.
Do not merge multiple papers into one only detailed report.
The report must be at least as rich as the examples under E:/论文调研/reports/accepted_papers/*.md, and it must also include the extra graph-ready and direction-mining sections required by this stage.
Treat those accepted-paper examples as the minimum top-conference deepread bar for the main narrative, not just for the appendix.
This skill now includes the Research-Generative Paper Reading lens. Apply it as a mandatory overlay on top of all original paper_deep_reading_skill requirements.
For every paper, explicitly extract the paper's research equation:
\text{Important Setting} + \text{Broken Assumption} + \text{Borrowed Tool} + \text{New Constraint} + \text{Surrogate Mechanism}. ]
Also express the paper as:
[ A(P) \cap \neg C \cap T \cap M \Rightarrow Z \approx Y. ]
Where:
The report must answer: what unavailable mechanism did the authors replace, and how did they make the replacement feel inevitable?
Add a research-discovery reconstruction using:
\text{Valuable Field} + \text{Painful Assumption} + \text{Emerging Tool} + \text{Unserved Setting}. ]
For each central idea, identify:
Use phrases such as:
For each paper, reconstruct the story bridge:
[ \text{Challenge}_i \rightarrow \text{Failure Mode}_i \rightarrow \text{Design Principle}_i \rightarrow \text{Module}_i \rightarrow \text{Ablation}_i. ]
Create a table:
| Challenge | Failure mode | Design principle | Module | Evidence |
|---|
Then decide whether the method is a weak additive story:
[ M_1 + M_2 + M_3 ]
or a stronger closed-loop story:
[ M_1 \Rightarrow M_2 \Rightarrow M_3 \Rightarrow M_1. ]
Name the object that flows through the loop, such as pseudo-labels, generated data, graph consensus, uncertainty, prototypes, rewards, or memory.
For every major module, use this template in addition to the existing formula/module critique:
\text{Failure} + \text{Unavailable Ideal Solution} + \text{Available Proxy} + \text{Design Choice} + \text{Assumption} + \text{Risk}. ]
The report must identify:
Do not only list related work. For important citations, infer their narrative function:
| Citation cluster | Narrative function | Assumption inherited | How this paper modifies it |
|---|
Treat citations as permissions to make moves such as: field foundation, limitation evidence, method ancestor, neighboring-field transfer, strong baseline, benchmark protocol, implementation machinery, or negative contrast.
Read each experiment as:
\text{Claim} + \text{Counterfactual} + \text{Metric} + \text{Stress Condition}. ]
For every major experiment block, state:
Extract at least one reusable story template from the paper, especially:
[ A \text{ solves } P \text{ when } C; \quad T \Rightarrow \neg C; \quad M \text{ helps but requires } Y; \quad Y \notin T; \quad Z \approx Y. ]
Also look for:
For each major weakness, use:
\text{Evidence} + \text{Hidden Assumption}. ]
Then convert it into:
\text{Current Method} + \text{Violated Assumption} + \text{New Mechanism}. ]
Prioritize directions that expose scientific frontier questions rather than only engineering refinements.
This skill now includes the Teaching-Explainer Paper Reading lens. Apply it as a mandatory overlay on top of the original deep-reading and research-generative requirements.
The teaching layer distills public best practices from paper-reading seminars, reviewer guides, and paper-analysis skills. It should be used as a method, not as a citation substitute inside the paper report.
Before writing the final report, infer or ask for the likely audience when feasible. If the user does not specify, default to:
For each paper, the report must include an explicit explanation plan:
| Audience | What they already know | What they will likely miss | What must be explained first | How much math to show | What evidence will convince them |
|---|
Do not assume that a listener understands the paper's notation, task setting, benchmark culture, or hidden assumptions. Explain these before relying on them.
Run this loop in addition to the existing workflow:
The detailed report must include these sections in addition to all original sections:
面向讲解的受众画像与讲解目标3 层讲解摘要:30 秒 / 3 分钟 / 10 分钟讲解主线 Story Spine听众先修知识与概念铺垫公式 / 图表 / 实验的讲解脚本板书推导与小例子演示角色扮演式讨论问题可能被问到的问题与回答证据易误解点与纠偏PPT / 分享稿结构草案听众可带走的三句话These sections may appear near the end of the main report before the plain-language story summary, or inside a clearly marked teaching appendix, but they must be present in the authoritative report.
For each paper, produce a compact teaching spine:
Before this paper:
The field could do X under assumption C.
Pain:
In setting T, C fails, so X no longer works.
Tempting path:
Method family M almost helps, but it requires unavailable mechanism Y.
Key insight:
The paper constructs surrogate Z for Y.
Method:
Z is implemented through modules M1, M2, ...
Evidence:
Experiments/figures/theory show Z helps under stress condition S.
Caveat:
Z still depends on hidden assumption H.
Teaching punchline:
The paper's contribution is not just a module, but a replacement story: Y is unavailable, so Z is made useful.
This story spine must remain consistent with the research equation:
A(P) ∩ ¬C ∩ T ∩ M ⇒ Z ≈ Y
Provide explanations at three levels:
The report must explicitly state what is lost when moving from layer 2 or 3 down to layer 1.
The report must not be a generic summary. Explain the paper to the level where a reader could reproduce the core method, defend the claims in a review discussion, and teach the logic to another researcher.
Use this knowledge-dependency order whenever it is feasible:
For every key concept, use the sequence:
直觉 -> 数学公式或 formal definition -> 具体例子 -> 局限 / failure case
For every complex module, explicitly state:
For missing experimental or implementation details, do not invent. Explicitly list gaps such as hardware, runtime, memory, optimizer, learning rate, seeds, early stopping, hyperparameter search, preprocessing, data split, label definition, baseline source, whether baselines were rerun or reconstructed, code availability, and compute budget.
The experiment section must be unusually concrete. Include dataset scale, split protocol, label definition, baseline genealogy, baseline rerun/reconstruction status, metric meanings, main numeric patterns, exceptions to the headline result, ablation purpose, qualitative evidence, and reproduction risk.
End the technical explanation with at least one complete numeric toy example that connects training and inference: define tiny inputs, show module outputs or score/loss updates, state which parameters change during training, and then run the learned logic on one inference example. If the paper lacks enough detail for a faithful numeric example, build a clearly marked toy example and state exactly which parts are simplified.
This skill supports a second phase for generating presentation-friendly visual storyboards, but the first execution of the skill must never generate images. The first execution is a text-only deep-reading delivery.
When the user invokes this skill for a paper, first respond with a concise plan, then produce only the complete authoritative deep-reading report and required delivery/status information. This first run must include:
使用这个skill,根据状态,执行第1步:生成背景、旧方法缺陷、论文问题与灵感来源的多张连续的卡通图。生成多张连续的卡通图.使用这个skill,根据状态,告知下一步应该问什么。Assume the conversation may be stateless. In every text-only status or planning turn, remind the user to include the skill name and current state in the next command. Use wording like:
如果开启新会话或上下文丢失,请说:
使用这个skill,根据状态,执行第X步:生成多张连续的卡通图,...。如果不知道下一步怎么问,可以说:使用这个skill,根据状态,告知下一步应该问什么。
Before starting any image-generation phase, include this guidance in a text-only planning/status turn, not in the same turn as image generation:
imagegen skill when it is available. If imagegen is unavailable or insufficient, call ChatGPT Images 2.0 API or another available image-generation API to produce the images; do not try to replace the requested cartoon storyboard with SVG-only output.After Step 0 is complete, offer a multi-step image plan. The user may request one step at a time. Each step should generate a coherent set of multiple separate 16:9 images in the same visual identity: cartoon-comic academic infographic style, consistent narrator character, consistent title numbering, consistent color palette, clear panels, actual paper data where relevant, camera continuity, and logical continuity.
Recommended default steps:
| Step | Name | Purpose | Typical image count | Must include | Must avoid |
|---|---|---|---|---|---|
| Step 1 | 背景 / 旧方法缺陷 / 问题与灵感 | Explain why the problem matters before the algorithm | 4-6 | continuous event streams, snapshot limitations, feature scarcity, context mismatch, ambiguous boundary, paper motivation | algorithm internals and experiment results |
| Step 2 | 算法整体流程与模块 | Explain BAD from input event to score | 5-7 | current edge (u,v,t), recent-neighbor retrieval, Encoding Layer, Pairwise Compatibility Modeling, Boundary-based Anomaly Detector, training vs inference | final performance results |
| Step 3 | 实验部分 | Explain datasets, baselines, metrics, protocols, results, ablations, qualitative plots, reproducibility caveats | 5-7 | Table 1 numbers, 12 baselines, AUC/AP explanation, Table 2 key results, Table 3 ablation, Figure 2/3 qualitative explanation, missing GPU/runtime details | unsupported claims about hardware/time |
| Step 4 | 关键局限与答辩质疑 | Prepare defense visuals for limitations and likely reviewer questions | 3-5 | score/boundary caveat, training contamination, cold-start, reproducibility, AP caveat, fairness of baseline protocol | inventing reviewer comments not present in sources |
| Step 5 | 未来方向与创新图谱 | Turn the paper into research-direction visuals | 3-5 | hidden assumptions, violated settings, robust flow, inductive compatibility, online conformal boundary, causal/counterfactual compatibility | presenting future work as already validated |
| Step 6 | 汇报封面 / 总结 / Q&A backup | Presentation packaging | 2-4 | cover image, final takeaway image, Q&A map | adding new technical claims |
When the user asks to execute a visual step:
imagegen skill when available; otherwise use ChatGPT Images 2.0 API or another approved image API.At the end of Step 0 or any text-only status turn, provide concise examples:
使用这个skill,根据状态,执行第1步:生成背景、旧方法缺陷、论文问题与灵感来源的多张连续的卡通图。使用这个skill,根据状态,执行第2步:生成算法整体流程与各模块解释的多张连续的卡通图。使用这个skill,根据状态,执行第3步:生成实验部分的多张连续的卡通图,包含数据集、baseline、AUC/AP、主结果、消融和复现缺口。使用这个skill,根据状态,执行第4步:生成局限性与答辩质疑的多张连续的卡通图。使用这个skill,根据状态,执行第5步:生成未来方向和创新图谱的多张连续的卡通图。使用这个skill,根据状态,告知下一步应该问什么。For every preserved key formula, provide:
| Formula | What problem it solves | Term-by-term meaning | How to say it aloud | Where it appears in the algorithm | What can go wrong |
|---|
Also include one beginner-friendly concrete example for at least one central formula or update rule.
For every important figure/table already analyzed by the deep-read stage, add a teaching script:
| Visual | First thing to point at | Listener confusion risk | Claim supported | What not to overclaim | Suggested verbal explanation |
|---|
If a figure is visually persuasive but evidentially weak, teach that distinction explicitly.
Turn each experiment block into a question-answer unit:
Question the experiment answers:
What a skeptical listener might ask:
Setup / dataset / baseline / metric:
Expected result if the paper's claim is true:
Actual result:
What it proves:
What it does not prove:
How to explain the table or curve in one sentence:
For each paper, create role-based discussion prompts:
| Role | What this role checks | Required output |
|---|---|---|
| Archaeologist | prior-work ancestry and what is truly new | one older ancestor, one inherited assumption, one modification |
| Bug Hunter | rigor, correctness, reproducibility, clarity | three attack questions and one likely author response |
| Researcher | future directions and test-of-time value | two follow-up projects, one high-risk boundary idea |
| Industry Practitioner | deployability and cost-benefit | adoption conditions, complexity cost, practical failure mode |
| Social Impact Assessor | risks, misuse, bias, privacy, safety | one positive impact, one omitted negative impact |
| Author Defender | rebuttal and best-paper pitch | strongest acceptance argument and weakest vulnerable claim |
| Teacher | whether a non-specialist can follow | analogy, prerequisite list, teachback questions |
Do not create a separate final report that competes with the authoritative detailed report. If a talk or slide blueprint is needed, make it a derivative artifact that cites the authoritative report as its source.
Minimum slide blueprint:
| Slide | Title | One-sentence point | Visual / formula | Speaker notes | Transition | Likely question |
|---|
Recommended default structure:
Prepare likely questions from three audiences:
For each answer, point back to specific evidence from the paper or state that the paper does not provide enough evidence.
Every teaching report must include:
End the teaching section with 8-12 questions a listener should be able to answer after the explanation. Include at least:
When useful, create derivative sidecar artifacts under:
generated/teaching/<paper-slug>/teaching_outline_cn.mdgenerated/teaching/<paper-slug>/slide_blueprint_cn.jsongenerated/teaching/<paper-slug>/qa_bank_cn.jsongenerated/teaching/<paper-slug>/role_play_discussion_pack_cn.mdThese sidecars must not replace the authoritative detailed report. They should cite or reference the sections of the authoritative report they derive from.
The following requirements are mandatory even when they are not strongly emphasized by the original paper:
Title interpretation rule: explicitly interpret the paper title term by term. Explain what each keyword means, why the title is phrased that way, and how the title maps onto the actual method, setting, and claims in the paper.
Mentioned-paper expansion rule: do not only list cited papers. For the papers that the target paper repeatedly discusses, compare what they solved, what they still left open, and why the current paper needed to move beyond them.
Multi-layer scientific-problem rule: explain both the paper-level problem and the upper problem ladder. Prefer continuous ladders such as direction-native -> parent-field -> broader AI/ML, and also discuss algorithm-lineage and bottleneck-derived upper problems when defensible.
Claim-audit rule: extract the paper's innovation points and core claims one by one, then audit whether each claim is supported by theory, experiments, ablations, visual evidence, reviewer discussion, or only by suggestive narrative. Say explicitly when support is weak, indirect, or missing.
Research-thought reconstruction rule: reconstruct the likely reasoning path that led the authors from observed pain points and prior work to the final design. Also infer what papers, methods, or upper-level scientific problems may have inspired the idea. Separate evidence-backed inference from speculation.
Author-subjectivity linkage rule: when discussing the idea, theory, proof choices, algorithm, or modules, explicitly note where the design may reflect the authors' subjective judgments, tastes, priors, heuristics, engineering preferences, or research style. Distinguish objective necessity from author-specific choice whenever the paper gives enough evidence.
Paper-grounded specificity rule: do not stay at a generic or abstract summary level. Stay close to the actual paper by naming the concrete modules, symbols, assumptions, losses, theorem objects, datasets, baselines, figures, and experimental findings that the authors actually use.
Related-work relation rule: explain the main related-work clusters and state the exact relation between each cluster and the current paper: inherited, contrasted, generalized, specialized, hybridized, or problem-shifted.
Symbols-and-concepts rule: explain symbols, assumptions, operators, and problem-specific concepts in beginner-friendly language before relying on them in later sections.
Formula-preservation rule: do not drop, compress away, or replace key equations with prose-only summaries. Preserve the paper's core objective functions, update rules, constraints, estimators, bounds, and theorem statements in readable form, then explain each formula term by term and state what role the formula plays in the method.
Proof-to-practice mapping rule: if the paper contains theory or proofs, explicitly map theorem assumptions, intermediate claims, and conclusions to the practical algorithm, implementation steps, or system behavior. State whether the implementation is exactly matched to the proved object, only approximately aligned, or only motivated by the theory.
Theory-purpose rule: explain why each proof or theoretical claim was included by the authors, what reviewer doubt or scientific concern it addresses, what practical meaning it has, and where the theory stops being faithful to the real implementation.
Module-and-equation critique rule: do not stop at restating the design. For each central formula, module, and modeling assumption, judge whether it may be brittle, under-justified, overly heuristic, computationally expensive, hard to optimize, poorly identified, or mismatched to the claimed scientific goal. Then discuss concrete improvement room, alternative formulations, and what trade-offs those changes would create.
Concrete algorithm walkthrough rule: the algorithm section must not stop at equations. Give a step-by-step procedure and at least one small concrete example that instantiates the states, inputs, outputs, and updates.
Experiment genealogy rule: when baselines are compared in the experiments, explain how those baselines relate to the algorithms and literature introduced earlier in the paper, especially the Introduction and Related Work sections.
Experiment-purpose rule: for every major experiment block, state what question it is trying to answer, what reviewer doubt it addresses, and what kind of evidence it produces. Also point out interesting, surprising, or controversial empirical phenomena.
Reviewer-lens enrichment rule: enrich the deep-reading report with extra reviewer lenses inspired by strong GitHub paper-review / peer-review skills. At minimum, explicitly audit novelty, significance, technical soundness, methodological rigor, reproducibility, results-claims alignment, missing baselines or controls, figure/table clarity, limitation honesty, and venue-fit or acceptance logic when relevant.
LaTeX-figure explanation rule: when a LaTeX source is available, do not ignore figures just because the paper is not being read from a rendered PDF. Reconstruct important figures from figure environments, captions, labels, referenced text, and included image paths when possible, then explain what each figure is trying to show, how it supports the argument, and what it leaves unclear.
PDF-figure explanation rule: when a PDF source is available, important figures must also be explicitly interpreted rather than only glanced at. Explain each key architecture figure, pipeline figure, qualitative example, and experimental plot in terms of purpose, visual structure, figure-to-claim mapping, and what remains unclear, potentially misleading, or insufficiently supported.
Experiment-chart consistency rule: in the experiment section, explicitly explain important tables, charts, curves, qualitative figures, and numeric comparisons. State which claim each visual or data block supports, whether the evidence really matches the stated claim, and where there is partial support, contradiction, or unexplained mismatch. Analyze plausible reasons for those mismatches instead of smoothing them away.
Innovation-type rule: judge whether the work is mainly incremental, cross-domain, boundary-breaking, or a mixture. Explain why.
Boundary-crossing rule: discuss whether the paper truly crosses a disciplinary or subfield boundary, or mainly recombines ideas inside one field.
Limitation rule: explicitly discuss the paper's weaknesses, unresolved assumptions, failure modes, scope limits, and risks of overclaim. Distinguish between minor limitations and structural limitations that cap the paper's impact.
Scientific-frontier direction rule: when proposing future work, distinguish ordinary follow-up engineering from directions that could actually push scientific boundaries. State what new scientific question, cross-field bridge, or conceptual shift would be required for a boundary-pushing follow-up.
New-direction rule: speculate about follow-up directions and new innovation hooks, including cross-domain transfer, lateral migration of the method, native extensions of the paper's own line, and higher-risk directions that may open a new boundary rather than only improve a benchmark.
Story-summary rule: end with one simple and vivid story that a non-specialist could understand without equations. That means the authoritative report must explicitly cover:
论文信息
论文标题解读
这篇论文真正解决的是什么
论文中提到的其他论文做了什么、留下了什么空白、与本文是什么关系
核心方法到底在干什么
公式保留与逐式解释
PDF 版本关键图解释
理论、证明与实现步骤对照
具体公式、模块与设计假设的不足及可改进空间
创新点、核心主张与证据逐条核对
这篇论文为什么重要 / 为什么值得被接收
实验是如何被设计出来的
倒推作者怎么想到这个 idea
idea / 理论证明 / 具体算法或模块里哪些部分更像作者的主观选择、经验判断或研究风格
报告必须结合论文本身的具体模块、公式、图表、实验与措辞,不能停留在过于抽象的泛化总结
审稿人最关注什么
额外审稿视角审计(借鉴 GitHub 热门审稿 skill 的 reviewer 关注点)
作者是怎么回复的
审稿人是否认同作者回复
审稿意见回复覆盖核对
强点 / 弱点 / 不足
作者团队近年的相关延续
未来边界
创新类型判断(增量 / 交叉 / 边界突破)
对选题的直接启示
可能的新研究方向或新创新点(尤其是可能推动科学边界的方向)
通俗生动故事总结
参考来源
再加上本 stage 需要的结构化补充附录
其中结构化附录里必须显式包含“公式保留与逐式解释”“理论、证明与实现步骤对照”以及“具体公式、模块与设计假设的不足及可改进空间”
The detailed report length is not capped. Prefer completeness over brevity when the paper warrants it.
In ChatGPT Projects, the default input should be one uploaded paper-collection zip in Project sources.
That zip should contain the paper PDFs and metadata bundle produced by the collection stage. Do not assume a local filesystem paper folder exists.
If a clean local scaffold is needed before building the upload bundle, initialize it with:
scripts/init_paper_deep_reading_scaffold.pyscripts/build_paper_deep_reading_bundle.pyscripts/validate_detailed_report_structure.pyThat scaffold should create:
metadata/query_spec.jsonmetadata/paper_batch_manifest.jsonmetadata/delivery_bundle_manifest.jsonmetadata/routing_status.jsonmetadata/project_directory_index.jsonreports/project_directory_index.mdreports/stage_delivery_handoff.mdreports/per_paper/<paper-slug>/<paper-slug>_detailed_cn.mdWork in stages. If the next step depends on files not yet present in the current Project sources, stop and ask the user to upload them first.
This especially includes:
When extracting upper scientific problems, do not rely only on the paper's own phrasing. Build a progressive problem ladder instead of making a large abstraction jump. Prefer:
Example:
PFL -> FL -> AI/MLDo not jump directly from PFL to a remote AI/ML abstraction if FL is the scientifically meaningful middle layer.
Only skip the middle layer when no defensible parent-field layer exists, and say so explicitly.
Also infer the upper ladder from:
Keep those three paths separate in the machine-readable outputs so a downstream graph page can show:
For each path, preserve the nearest parent-field layer before the broader ML/AI layer whenever the evidence supports it.
If the paper is from ICLR, use the paper title and year to find the matching OpenReview forum before writing the review/rebuttal section. Use that forum as the canonical source for reviewer concerns, author replies, and decision context.
If the paper came from a previous collection stage, prefer consuming the already downloaded local OpenReview bundle from the project tree instead of re-searching first. The detailed report should explicitly use that local bundle when it exists.
If an ICLR paper still lacks openreview_forum_url, report that gap explicitly in the machine-readable output and in the human-facing report instead of silently continuing as if review context were complete.
If the paper is not from ICLR, do not force a review/rebuttal section by default unless the user explicitly asks for external peer-review context and a reliable public source exists.
For the proposed method, prefer extracting more than a short summary. Capture:
Also analyze how the paper narrows the gap between motivation and final algorithm. Explicitly look for:
When summarizing the proposed method, do all of the following when the source materials allow it:
For experiments, also require:
For each one, state what claim it is trying to support and whether it really makes the final method more convincing.
Example:
LoRA, check whether that points upward to feature alignment, shared-subspace design, or interface-stable transferThis stage must not reopen external retrieval.
Do not, inside this stage:
Instead:
For human-facing reading output, keep one authoritative report file.
sources refresh bundle, prefer intermediate JSON only; do not include intermediate Markdown unless the user explicitly wants audit traces.ChatGPT canvas should prefer standard-library-safe scripts and source files already placed in sources.
If PDF rendering, OCR, or local figure extraction has already been done outside ChatGPT, ask the user to upload:
visual_manifest.jsonPreferred OCR preprocessing order before upload:
RapidOCRTesseract + pytesseractEasyOCRIf the local OCR chain is unavailable, upload the rendered page images anyway and let ChatGPT continue with PDF text layer, captions, and nearby blocks.
Do not keep OCR lines, OCR block parsing, or figure-analysis confidence in the default final artifacts unless the user explicitly wants OCR audit traces.
When a local preprocessing batch is already finished, only upload final artifacts into Project sources.
Do not upload probe or test-only OCR directories such as visuals_test or visuals_easyocr_probe*.
Prefer a final bundle that contains:
visual_manifest.jsonvisuals/pages/ directory if the Markdown still references those imagesBefore handing the bundle to the next stage, run:
scripts/validate_detailed_report_structure.pyscripts/build_paper_deep_reading_bundle.pyDo not assume ChatGPT can access local files outside Project sources.
workflow/01_request_sources.mdworkflow/02_extract_structure.mdworkflow/03_figure_and_table_analysis.mdworkflow/04_gap_mining_and_graph.mdworkflow/05_extended_argument_and_innovation_audit.mdworkflow/06_teaching_explainer_preparation.mdworkflow/07_talk_qa_rehearsal.mdworkflow/08_staged_cartoon_visual_storyboard.mdworkflow/09_storyboard_pdf_assembly.mdscripts/assemble_storyboard_pdf.pyschemas/paper_focus_spec.template.jsonschemas/detailed_report_contract.mdschemas/detailed_report_required_sections.jsonschemas/per_paper_output_layout.mdschemas/project_directory_index_spec.mdschemas/project_directory_annotations.template.jsonschemas/motivation_bridge_analysis.mdschemas/research_generative_overlay.mdschemas/teaching_explanation_overlay.mdschemas/teaching_explanation_spec.template.jsonschemas/external_best_practices_sources.mdschemas/routing_status_template.jsonschemas/sources_zip_layout.mdschemas/sources_refresh_templates.mdscripts/build_canvas_deepread_intermediate.pyscripts/init_paper_deep_reading_scaffold.pyscripts/build_paper_deep_reading_bundle.pyscripts/validate_detailed_report_structure.pyscripts/update_project_directory_index.pyscripts/update_routing_status.pyAt the end of every substantive reply, append:
Current StatusPossible User Inputs For Next StageDo not include a Recommended Next Skill section. When visual generation is the likely next action, include a user prompt that explicitly says 生成多张连续的卡通图.
metadata/paper_batch_manifest.json. Do not select-read only part of the batch.ICLR, consume the already-downloaded local OpenReview review / rebuttal bundle instead of skipping review context.scripts/validate_detailed_report_structure.py. That validator should fail if any paper in the batch still lacks its authoritative report.*_detailed_cn.md*_detailed_cn.pdfintermediate/*.jsongenerated/teaching/ when generatedsourcesmetadata/query_spec.jsonmetadata/paper_batch_manifest.jsonmetadata/delivery_bundle_manifest.jsonmetadata/routing_status.jsonmetadata/project_directory_index.jsonreports/project_directory_index.mdreports/stage_delivery_handoff.mdsourcesresearch_subfield or a confirmed direction labelbudget / hardware information before ranking directionspaper_idpaper_title
key_artifactsblockersmetadata/project_directory_index.json before searching the tree ad hoc