<?xml version="1.0" encoding="utf-8" ?><rss version="2.0"><channel><title>Bing: How to Initialize a Character in Java</title><link>http://www.bing.com:80/search?q=How+to+Initialize+a+Character+in+Java</link><description>Search results</description><image><url>http://www.bing.com:80/s/a/rsslogo.gif</url><title>How to Initialize a Character in Java</title><link>http://www.bing.com:80/search?q=How+to+Initialize+a+Character+in+Java</link></image><copyright>Copyright © 2026 Microsoft. All rights reserved. These XML results may not be used, reproduced or transmitted in any manner or for any purpose other than rendering Bing results within an RSS aggregator for your personal, non-commercial use. Any other use of these results requires express written permission from Microsoft Corporation. By accessing this web page or using these results in any manner whatsoever, you agree to be bound by the foregoing restrictions.</copyright><item><title>Clever: A Curated Benchmark for Formally Verified Code Generation</title><link>https://openreview.net/attachment?id=pqNFDA2TFm&amp;name=pdf</link><description>We introduce CLEVER, the first curated benchmark for evaluating the generation of specifications and formally verified code in Lean. The benchmark comprises of 161 programming problems; it evaluates both formal speci-fication generation and implementation synthesis from natural language, requiring formal correctness proofs for both.</description><pubDate>Sun, 12 Apr 2026 06:10:00 GMT</pubDate></item><item><title>CLEVER: A Curated Benchmark for Formally Verified Code Generation</title><link>https://openreview.net/forum?id=pqNFDA2TFm</link><description>TL;DR: We introduce CLEVER, a hand-curated benchmark for verified code generation in Lean. It requires full formal specs and proofs. No few-shot method solves all stages, making it a strong testbed for synthesis and formal reasoning.</description><pubDate>Fri, 10 Apr 2026 21:00:00 GMT</pubDate></item><item><title>CLEVER: A Curated Benchmark for Formally Verified Code Generation</title><link>https://openreview.net/forum?id=IbOacMF5qd</link><description>This paper introduces CLEVER, a benchmark dataset designed to evaluate LLMs on formally verified code generation. It consists of 161 carefully crafted Lean specifications derived from programming problems in the existing HumanEval dataset.</description><pubDate>Sat, 11 Apr 2026 17:52:00 GMT</pubDate></item><item><title>The Clever Hans Mirage: A Comprehensive Survey on Spurious...</title><link>https://openreview.net/forum?id=kIuqPmS1b1</link><description>This survey on spurious correlations uses the Clever Hans metaphor to motivate the problem, formalizes a group-based setup g=(y,a) with core metrics (worst-group, average-group, bias-conflicting), and explains why models latch onto shortcuts (simplicity bias, training dynamics).</description><pubDate>Sat, 11 Apr 2026 17:52:00 GMT</pubDate></item><item><title>Counterfactual Debiasing for Fact Verification</title><link>https://openreview.net/pdf?id=BddNTCq65yq</link><description>579 In this paper, we have proposed a novel counter- factual framework CLEVER for debiasing fact- checking models. Unlike existing works, CLEVER is augmentation-free and mitigates biases on infer- ence stage. In CLEVER, the claim-evidence fusion model and the claim-only model are independently trained to capture the corresponding information.</description><pubDate>Thu, 09 Apr 2026 23:17:00 GMT</pubDate></item><item><title>STAIR: Improving Safety Alignment with Introspective Reasoning</title><link>https://openreview.net/forum?id=aHzPGyUhZa</link><description>One common approach is training models to refuse unsafe queries, but this strategy can be vulnerable to clever prompts, often referred to as jailbreak attacks, which can trick the AI into providing harmful responses. Our method, STAIR (SafeTy Alignment with Introspective Reasoning), guides models to think more carefully before responding.</description><pubDate>Sat, 11 Apr 2026 17:52:00 GMT</pubDate></item><item><title>Evaluating the Robustness of Neural Networks: An Extreme Value...</title><link>https://openreview.net/forum?id=BkUHlMZ0b</link><description>Our analysis yields a novel robustness metric called CLEVER, which is short for Cross Lipschitz Extreme Value for nEtwork Robustness. The proposed CLEVER score is attack-agnostic and is computationally feasible for large neural networks.</description><pubDate>Sat, 11 Apr 2026 17:52:00 GMT</pubDate></item><item><title>On the Planning Abilities of Large Language Models : A Critical ...</title><link>https://openreview.net/pdf?id=X6dEqXIsEW</link><description>While, as we mentioned earlier, there can be thorny “clever hans” issues about humans prompting LLMs, an automated verifier mechanically backprompting the LLM doesn’t suffer from these. We tested this setup on a subset of the failed instances in the one-shot natural language prompt configuration using GPT-4, given its larger context window.</description><pubDate>Fri, 10 Apr 2026 15:30:00 GMT</pubDate></item><item><title>Measuring Mathematical Problem Solving With the MATH Dataset</title><link>https://openreview.net/forum?id=7Bywt2mQsCe</link><description>Abstract: Many intellectual endeavors require mathematical problem solving, but this skill remains beyond the capabilities of computers. To measure this ability in machine learning models, we introduce MATH, a new dataset of 12,500 challenging competition mathematics problems. Each problem in MATH has a full step-by-step solution which can be used to teach models to generate answer derivations ...</description><pubDate>Sat, 11 Apr 2026 17:31:00 GMT</pubDate></item><item><title>Submissions | OpenReview</title><link>https://openreview.net/submissions?page=63&amp;venue=ICLR.cc%2F2025%2FConference</link><description>Promoting openness in scientific communication and the peer-review process</description><pubDate>Tue, 14 Apr 2026 20:12:00 GMT</pubDate></item></channel></rss>