¥
立即购买

大学生英语写作专家

246 浏览
24 试用
5 购买
Dec 6, 2025更新

智能搞定英语写作,从论文到邮件,地道表达轻松上手,写作水平飙升!

Balancing Writing Quality and Academic Integrity: A Quasi-Experimental Evaluation of Generative AI in a First-Year Composition Course

Abstract

Generative artificial intelligence (AI) is rapidly entering university writing pedagogy, promising improved feedback, language support, and productivity. Yet its adoption raises concerns about academic integrity, authorship, and the erosion of core writing competencies. This study reports a quasi-experimental evaluation of structured generative AI support in two sections of a first-year composition course (N = 126). The intervention combined explicit AI-use boundaries, disclosure requirements, prompt-and-output logging, and revision-focused scaffolds. Outcomes included rubric-based writing quality, text similarity, confirmed integrity breaches, and process analytics. Controlling for pretest scores, the AI-supported section achieved significantly greater gains in overall writing quality (d = 0.58) without a statistically significant increase in similarity indices or integrity violations compared to the control section. Effects were strongest for lower-proficiency writers and were partially mediated by the number of AI–revision cycles. Findings suggest that, when bounded and transparent, generative AI can enhance writing quality while maintaining academic integrity. The study proposes a principled framework for permissible uses, process-oriented assessment, and verification mechanisms. Limitations include single-institution scope, short duration, and imperfect detection of undisclosed AI. Implications for policy, curriculum design, and future research are discussed.

Keywords

  • generative AI
  • academic writing
  • academic integrity
  • quasi-experiment
  • higher education

Introduction

Generative AI tools capable of synthesizing text and providing instant feedback are transforming the landscape of university writing instruction. Their potential to scaffold outlining, argument development, and language polishing is counterbalanced by concerns over plagiarism, undisclosed authorship, and over-reliance that may weaken students’ independent writing skills. Despite growing institutional guidance, empirical evidence quantifying the benefits and risks of generative AI in authentic writing courses remains limited.

This study addresses that gap by evaluating a structured, transparent, and revision-centered integration of generative AI in a first-year composition course. We compare writing quality and integrity outcomes between an AI-supported section and a control section following a quasi-experimental pretest–posttest design. We position writing as a socio-cognitive process in which tools can extend, but not replace, learner agency and instructor feedback.

Research questions:

  • RQ1: Does structured generative AI support improve writing quality relative to traditional instruction?
  • RQ2: Does structured AI support alter academic integrity outcomes (similarity indices, confirmed violations) compared with a control condition?
  • RQ3: Do effects vary by baseline writing proficiency?
  • RQ4: Which patterns of AI use (e.g., number of AI–revision cycles) are associated with gains without increasing integrity risk?

Argument: With clearly defined boundaries, process transparency, and assessment anchored in students’ iterative work, generative AI can produce meaningful gains in writing quality without compromising academic integrity.

Literature Review

Early classroom implementations report that AI-assisted feedback can enhance idea generation, organization, and language accuracy, especially for novice writers (Author & Author, Year). However, studies also caution that unbounded use can blur authorship and increase reliance on AI-generated phrasing (Author, Year). The literature is converging on the importance of scaffolding: guiding prompt design, emphasizing revision rather than substitution, and requiring reflective commentary on AI outputs (Author, Year).

Three representative strands are notable. First, controlled studies of AI-supported drafting show moderate improvements in coherence and argument structure when AI is used to refine outlines and thesis statements rather than to produce full drafts (Author & Author, Year). Second, integrity-focused research highlights the limits of AI detection and the risk of false positives, recommending process evidence (drafts, logs, oral defenses) over product-only policing (Author, Year). Third, design-oriented work demonstrates that combining AI with formative feedback loops and metacognitive reflection yields the largest learning gains, particularly for lower-proficiency learners (Author, Year).

Despite these advances, two gaps persist. There is limited quasi-experimental evidence linking specific AI-use boundaries to both quality and documented integrity outcomes, and few studies triangulate rubric scores, similarity metrics, and process analytics to separate genuine learning from mere textual polish. This study addresses both gaps by embedding AI within a transparent, revision-centered pedagogy and by analyzing outcomes at multiple levels.

Methodology

Design

We employed a quasi-experimental pretest–posttest design with non-equivalent groups (two intact course sections) over a 10-week term. One section implemented structured generative AI support (AI condition); the other used traditional instruction without AI (control). Pretest writing performance served as a covariate in subsequent analyses.

Participants

Participants were 126 first-year undergraduates enrolled in a required composition course at a large public university (AI: n = 64; control: n = 62). Students represented varied majors and linguistic backgrounds; 41% reported English as an additional language. No significant differences emerged between sections on pretest writing scores or demographic variables.

Instruments

  • Writing quality rubric: A validated analytic rubric (20-point scale) assessing argumentation, organization, evidence integration, language accuracy, and audience awareness. Two trained raters scored anonymized samples; interrater reliability was acceptable (ICC = .86).
  • Academic integrity indicators: Text similarity indices generated via an institutional plagiarism-detection system; confirmed violations included plagiarism or undisclosed AI use substantiated by document forensics (metadata), prompt/output logs (AI condition), and instructor review.
  • Process analytics: Counts of AI–revision cycles, prompt types (brainstorming, outlining, micro-editing), and time-on-task captured via a learning platform plugin (AI condition).
  • Surveys: Brief, non-graded questionnaires on perceived learning and effort (both conditions), used descriptively.

Procedure

Week 1: Both sections completed a timed diagnostic essay (pretest). The AI condition received a 60-minute workshop on permissible AI uses, disclosure requirements, and prompt engineering for revision (not drafting). Students signed an AI-use statement and learned to log prompts and outputs within the platform.

Weeks 2–9: Both sections completed two major essays (argumentative and synthesis). The control section followed standard drafting with peer review and instructor feedback. The AI section used AI only for:

  • Idea generation (brainstorming lists, research questions),
  • Macro-structure (outline refinement, thesis calibration),
  • Micro-editing (grammar, style suggestions with rationales).

Full-draft generation by AI was prohibited. All AI interactions were logged automatically and discussed in brief reflective memos attached to submissions. Both sections received equivalent instructor feedback cycles.

Week 10: A posttest essay under classroom conditions (no AI) assessed transfer. All final submissions underwent similarity checking; suspected violations were reviewed by a blinded academic integrity panel.

Data Analysis

We conducted ANCOVA on posttest rubric scores with pretest scores as covariates. Domain-level effects (e.g., organization) were analyzed similarly. Logistic regression modeled odds of confirmed integrity violations. Moderation by baseline proficiency (terciles by pretest score) and mediation by AI–revision cycles (AI condition only) were examined using interaction terms and a simple product-of-coefficients approach. Alpha was set at .05; effect sizes (d, partial η², OR) and 95% confidence intervals are reported.

Results

Writing Quality

Controlling for pretest performance, the AI condition outperformed the control on posttest overall writing quality, F(1, 121) = 18.41, p < .001, partial η² = .13. The adjusted mean difference corresponded to d = 0.58, 95% CI [0.31, 0.85]. Domain-specific analyses showed:

  • Organization and coherence: d = 0.62, p < .001
  • Evidence integration and citation: d = 0.50, p = .002
  • Language accuracy and style: d = 0.44, p = .006
  • Audience awareness: d = 0.33, p = .032

Academic Integrity Outcomes

Average similarity indices did not differ significantly between groups: AI M = 13.5% (SD = 6.4), control M = 12.8% (SD = 6.1), t(124) = 0.68, p = .50. Confirmed integrity violations were low and statistically comparable: AI 3.1% (2/64) vs. control 4.8% (3/62), Fisher’s exact p = .68; logistic regression OR = 0.64, 95% CI [0.10, 3.96], p = .62. In the AI section, all students submitted prompt logs and disclosures; two cases involved inadequate paraphrasing of AI-suggested phrasing. In the control section, three cases involved unattributed copying from online sources.

Moderation and Mediation

Baseline proficiency moderated treatment effects, F(2, 119) = 6.94, p = .010. The effect was largest for the lowest proficiency tercile (d = 0.82), moderate for the middle tercile (d = 0.49), and small for the highest tercile (d = 0.28). Within the AI condition, the number of AI–revision cycles predicted greater gains, b = 0.12 points per cycle (SE = 0.04), p = .002; a simple mediation model indicated partial mediation of the treatment effect by revision cycles (indirect effect = 0.31, 95% CI [0.11, 0.57]). Time-on-task was slightly higher in the AI condition (+18 minutes per essay on average; p = .047), suggesting augmented—not reduced—effort.

Transfer

On the in-class, no-AI posttest, the AI section maintained advantages in organization and evidence use, supporting transfer of strategy rather than dependence on AI outputs.

Discussion

Interpreting the Findings

The intervention yielded moderate improvements in writing quality without elevating integrity risks, addressing RQ1 and RQ2. Gains concentrated in higher-order features—organization and evidence use—consistent with literature emphasizing AI’s value for macro-structural support when paired with human judgment (Author & Author, Year). The moderation finding answers RQ3: lower-proficiency writers benefited most, likely because AI scaffolds reduced cognitive load and made revision moves more visible (Author, Year). Mediation by AI–revision cycles answers RQ4 and aligns with a process-oriented account: it is not AI per se, but structured iterations with reflection, that drive learning (Author, Year).

Boundaries for Generative AI Use

Evidence from this study supports the following boundaries:

  • Permissible: brainstorming, outline and thesis refinement, query-driven feedback on argument development, and micro-editing with rationales.
  • Prohibited: whole-draft generation, automated citation fabrication, and any substitution that obscures student authorship.
  • Required: full disclosure of AI use, prompt/output logs, and brief reflective memos explaining what was accepted, modified, or rejected and why.

These boundaries operationalize a “human-in-the-loop” model that preserves authorship while leveraging AI for formative support.

Academic Integrity Strategies

A multi-layer integrity framework proved effective:

  • Transparency: mandatory disclosures and logs reduced ambiguity around authorship and provided evidence for review.
  • Process-based assessment: drafts, annotations, and short oral defenses prioritized students’ reasoning and revision decisions over product-only evaluation.
  • Targeted detection: similarity checking flagged conventional plagiarism; instructors avoided over-reliance on AI detectors given reliability concerns (Author, Year).
  • Clear policy and instruction: early, specific guidance on acceptable use reduced inadvertent violations and reinforced shared norms.

Violation rates remained low and comparable across groups, suggesting that well-communicated boundaries and verifiable process evidence can maintain integrity while enabling AI-supported learning.

Limitations

The quasi-experimental design with intact sections limits causal claims relative to randomized trials. The setting—a single institution and course—constrains generalizability across disciplines. Detection of undisclosed AI remains imperfect, even with logs. Finally, the 10-week duration may underestimate longer-term effects, including potential over-reliance or normalization of poor practices if guardrails weaken.

Implications

For policy: codify permissible uses, mandate disclosure, and prefer process evidence over high-stakes AI detection. For pedagogy: design assignments that require drafts, metacognitive reflection, and verbal defense; teach prompt engineering for revision rather than generation. For equity: structured AI scaffolds may particularly support lower-proficiency and multilingual writers; institutions should ensure equitable access and training.

Conclusion

When integrated with clear boundaries, transparency, and a focus on revision, generative AI can produce meaningful gains in writing quality without compromising academic integrity. Our quasi-experimental evidence shows moderate improvements in higher-order writing features, stable integrity metrics, and stronger benefits for lower-proficiency students. Moving forward, research should test variations of scaffolded AI use across disciplines, examine long-term learning and transfer, and refine verification strategies that respect student privacy while safeguarding integrity. A principled, process-centered approach offers a viable path to balance innovation with the core values of academic writing.

References

Author, A. A., & Author, B. B. (Year). Generative AI as a scaffold for academic writing: Effects on organization and coherence. Journal of Writing Research, 15(2), 123–147. https://doi.org/10.0000/jwr.2023.0001

Author, C. C. (Year). Authorship, originality, and AI: Reframing academic integrity in higher education. Ethics and Education, 18(1), 45–62. https://doi.org/10.0000/ee.2024.0002

Author, D. D., Author, E. E., & Author, F. F. (Year). Designing AI-in-the-loop writing instruction: From prompt engineering to reflective revision. Computers & Education, 210, 104778. https://doi.org/10.0000/cae.2024.104778

Author, G. G., & Author, H. H. (Year). Beyond detection: Process evidence as a safeguard in the age of generative text. Assessment & Evaluation in Higher Education, 49(3), 367–382. https://doi.org/10.0000/aehe.2023.0003

Author, I. I. (Year). Quasi-experimental designs for educational interventions: Practical considerations. Educational Researcher, 52(4), 210–222. https://doi.org/10.0000/er.2023.0004

Author, J. J. (Year). Cognitive load, feedback loops, and AI-supported writing: A process perspective. Learning and Instruction, 85, 101724. https://doi.org/10.0000/li.2022.101724

Author, K. K., & Author, L. L. (Year). Validity and limits of text similarity as a proxy for plagiarism. Journal of Academic Integrity, 8(1), 1–19. https://doi.org/10.0000/jai.2021.0005

Author, M. M. (Year). Transparency and accountability frameworks for educational AI. International Review of Education Technology, 12(4), 299–318. https://doi.org/10.0000/iret.2024.0006

Mandating Sustainability Literacy: A Prerequisite for Responsible Scholarship

Why Every Undergraduate Must Learn to Navigate a Warming World

What if the most valuable skill we carry across the graduation stage is not a major-specific technique, but the capacity to understand and manage the systems that sustain our lives? We argue that universities should require sustainability literacy for all undergraduates. Not as ideology, not as an elective, but as literacy—an indispensable precondition for informed citizenship and employability in a volatile century.

Argument 1: Professional Readiness and Civic Responsibility

Sustainability literacy is a cross-cutting competency, comparable to numeracy or academic writing. Employers in engineering, finance, healthcare, and the arts increasingly expect us to interpret material footprints, evaluate climate risks, and apply evidence-based heuristics to reduce waste and cost. On campus, we have seen how quickly foundational concepts become practical: in a first-year seminar, we audited plug loads in a residence hall and, by enabling sleep modes and using smart power strips, reduced electricity consumption by 12% in one month. That small project taught us to measure, to model, to manage. Beyond the résumé, this literacy equips us as citizens to read a budget line, a carbon inventory, or a flood map with critical acuity. The result is not only competence, but credibility.

Argument 2: Systems Thinking Across Disciplines

Sustainability is not a niche; it is a lens that clarifies complexity. Through systems thinking—feedbacks, trade-offs, lifecycle analysis—we grasp how decisions reverberate across ecological, economic, and social dimensions. This interdisciplinary fluency is urgently needed wherever problems are multifaceted. Consider a cross-listed capstone in which we redesigned the campus shuttle schedule: by applying queuing models and equity criteria, we cut empty runs by 18% while improving late-night service for off-campus students. The project demanded synthesis—statistics and ethics, logistics and lived experience. Such synthesis is a catalyst for innovation precisely because it disciplines imagination with constraints. In short, sustainability literacy teaches us to connect dots others do not even see.

Argument 3: The Campus as a Living Laboratory

When we treat the university as a living lab, literacy becomes action and action yields tangible benefits. A dining hall pilot that combined weigh-station data with menu planning decreased plate waste by nearly one-third in six weeks, saving money and emissions. In teaching labs, we replaced single-pass cooling with recirculators, conserving over 100,000 liters of water in a semester. The library’s default switch to duplex printing cut paper use dramatically with zero loss of productivity. These are not abstract ideals; they are pragmatic interventions, grounded in measurement and iterative design. The campus thus becomes both classroom and case study, embedding sustainability in daily practice.

Objection and Rebuttal

Some will object that the curriculum is already overcrowded, or that not every student is an environmental major. Others worry that “sustainability” smuggles in ideology. We share the concern for academic rigor—and that is precisely why a mandate should be modest, flexible, and evidence-driven. A three-credit requirement or a suite of embedded modules can be tailored to disciplines: lifecycle costing for business, materials toxicity for studio arts, climate risk for urban studies, environmental ethics for philosophy. Assessment can be granular and transparent—can we interpret a footprint, critique a trade-off, communicate uncertainty? Far from indoctrination, this approach privileges testable claims over slogans, data over dogma. If we can teach statistical inference without prescribing conclusions, why not sustainability literacy with the same intellectual safeguards?

Conclusion: From Principle to Practice

Mandating sustainability literacy is not a panacea, but it is a start—measurable, equitable, and overdue. We propose three steps. First, we set university-wide learning outcomes—systems thinking, lifecycle reasoning, risk communication—and adopt them through faculty governance. Second, we pilot the requirement in three departments next semester, using the campus as a living lab and publishing results openly. Third, we create a micro-credential that recognizes mastery gained through courses and applied projects, incentivizing participation without extending time to degree. Let us operationalize our values with urgency and care. Let us not wait, not wonder, not waver. We can build a curriculum that matches the zeitgeist of our century—one that prepares us to think long, act now, and lead together.

Subject: Application for Full-Time Summer Research Internship (June–August) in Your Cognitive Psychology Lab

Dear Professor [Last Name],

I am writing to inquire about a full-time summer research internship (June–August) in your cognitive psychology lab.

I am a junior majoring in Psychology (GPA 3.7) with training in experimental design and proficiency in Python, R, and SPSS. My coursework and research assistantship have equipped me to design and implement behavioral experiments, manage datasets, and conduct rigorous statistical analyses aligned with contemporary cognitive science.

In the Attention and Memory Lab at [University], I built and deployed a Stroop and n-back battery in PsychoPy, coordinated data collection (N = 118), and executed preprocessing and mixed-effects analyses in R (lme4). The study produced a robust congruency effect (mean RT difference ≈ 72 ms; error rate +5.9%), and a Load × Congruency interaction consistent with the literature. I co-authored a poster for our university research symposium and helped preregister the design and analysis plan on OSF, ensuring methodological transparency.

In a follow-up project on decision-making under time pressure, I designed a drift-diffusion modeling pipeline in Python (pandas, NumPy) and SPSS for confirmatory checks. I reduced data-cleaning time by approximately 45% through automated QA scripts and implemented reproducible reports (R Markdown). Our pilot (N = 56) yielded parameter estimates that distinguished speed–accuracy trade-offs across conditions; I drafted the methods and results sections for an internal report and prepared figures with ggplot2.

I am particularly drawn to your group’s work on cognitive control and learning, and I see a strong fit between my skills and your lab’s use of tightly controlled behavioral paradigms, quantitative modeling, and open, reproducible workflows. I would value the opportunity to contribute to ongoing projects, assist with experiment programming and data analysis, and learn advanced methods your team employs.

I am available full-time from early June through late August (up to 40 hours/week). I can work on-site if feasible, and I am also prepared to contribute effectively in a hybrid or fully remote arrangement. My CV and unofficial transcript are attached for your review.

Thank you for considering my application. I would be grateful for the chance to discuss how I can support your lab this summer.

Sincerely,
[Your Full Name]
[University Name], B.A. in Psychology (Expected [Month Year])
Email: [your.email@university.edu]
Phone: [+country code–phone number]

Attachments:

  • Curriculum Vitae (CV)
  • Unofficial Transcript

示例详情

解决的问题

帮助大学生在任何英语写作场景下快速产出专业级成稿:精准理解题目与读者、自动匹配文体与语气、搭建清晰结构、逐段生成与深度润色,确保地道表达与课堂规范双达标;显著提升写作成绩与通过率,同时缩短写作时间,形成可复用的高分写作方法论;覆盖论文、课程作业、竞赛作文、日常邮件与申请文书等全场景,助你从草稿到定稿一步到位。

适用用户

本科生与研究生

使用提示词从选题拆解到成稿,快速完成课程论文、读书报告与期末作文;自动优化逻辑与语言,提升评分与导师认可。

留学申请者

生成并打磨个人陈述、简历与推荐信草稿;按院校偏好调整语气与故事结构,优化邮件沟通与后续跟进,提高录取竞争力。

英语竞赛选手

快速构思观点与框架,生成比赛作文与演讲稿;增强论证、引用与过渡语,提升现场表达与评委好感度。

特征总结

一键选择写作类型,轻松生成论文、作文、邮件与申请文书,提交即用
自动规划结构与段落,快速搭建引言—论证—结论框架,逻辑清晰可读
智能匹配读者与语气,按课程要求或院校偏好调整表达与修辞,提升说服力与可读性
自动语法检查与词汇升级,输出地道英语表达,避免中式痕迹与低级错误
按题目需求生成素材与论点,提供例证与过渡语,让文章更有力更连贯
支持中英混合输入,精准理解题意,按要求输出标准格式与清晰排版
内置原创性保障与引用规范提醒,帮助规避抄袭与格式风险,安心提交评分
多轮自动优化,可按导师反馈微调语气与结构,快速迭代到满意版本
专为大学场景打造,覆盖课程作业、竞赛作文、邮件沟通与留学申请
一键生成不同风格版本,快速比较并选择更契合评分标准与读者期待的稿件

如何使用购买的提示词模板

1. 直接在外部 Chat 应用中使用

将模板生成的提示词复制粘贴到您常用的 Chat 应用(如 ChatGPT、Claude 等),即可直接对话使用,无需额外开发。适合个人快速体验和轻量使用场景。

2. 发布为 API 接口调用

把提示词模板转化为 API,您的程序可任意修改模板参数,通过接口直接调用,轻松实现自动化与批量处理。适合开发者集成与业务系统嵌入。

3. 在 MCP Client 中配置使用

在 MCP client 中配置对应的 server 地址,让您的 AI 应用自动调用提示词模板。适合高级用户和团队协作,让提示词在不同 AI 工具间无缝衔接。

AI 提示词价格
¥30.00元
先用后买,用好了再付款,超安全!

您购买后可以获得什么

获得完整提示词模板
- 共 630 tokens
- 6 个可调节参数
{ 写作主题 } { 写作类型 } { 具体要求 } { 目标读者 } { 期望字数范围 } { 期望语言风格 }
获得社区贡献内容的使用权
- 精选社区优质案例,助您快速上手提示词
使用提示词兑换券,低至 ¥ 9.9
了解兑换券 →
限时半价

不要错过!

半价获取高级提示词-优惠即将到期

17
:
23
小时
:
59
分钟
:
59