id
int64 393k
2.82B
| repo
stringclasses 68
values | title
stringlengths 1
936
| body
stringlengths 0
256k
⌀ | labels
stringlengths 2
508
| priority
stringclasses 3
values | severity
stringclasses 3
values |
---|---|---|---|---|---|---|
2,796,110,878 | go | net: TestUDPServer/2 failures | ```
#!watchflakes
default <- pkg == "net" && test == "TestUDPServer/2"
```
Issue created automatically to collect these failures.
Example ([log](https://ci.chromium.org/b/8725537367198926609)):
=== RUN TestUDPServer/2
server_test.go:258: udp [::ffff:0.0.0.0]:0<-127.0.0.1
server_test.go:322: server: read udp [::]:56739: i/o timeout
server_test.go:315: client: read udp 127.0.0.1:50097: i/o timeout
--- FAIL: TestUDPServer/2 (3600.00s)
— [watchflakes](https://go.dev/wiki/Watchflakes)
| NeedsInvestigation | low | Critical |
2,796,113,577 | next.js | Prefetched PPR page segments stay suspended when navigated to after a rebuild | ### Link to the code that reproduces this issue
https://github.com/bradleydsmith/reproduction-next-prefetch-rebuild-ppr-suspended
### To Reproduce
1. Build the application with npm run build
2. Start the server with npm run start
3. Navigate to http://localhost:3000 and do not click anything or close the page
4. Stop the server
5. Run npm run build
6. Start the server again with npm run start
7. Click the Prefetched link on the page
### Current vs. Expected behavior
When following the steps the suspended segment stays suspended and does not get resolved and rendered. I expected the page to do an MPA navigation and reload due to detecting a change in the response build ID
### Provide environment information
```bash
Operating System:
Platform: darwin
Arch: x64
Version: Darwin Kernel Version 23.5.0: Wed May 1 20:09:52 PDT 2024; root:xnu-10063.121.3~5/RELEASE_X86_64
Available memory (MB): 8192
Available CPU cores: 8
Binaries:
Node: 20.9.0
npm: 10.1.0
Yarn: N/A
pnpm: 9.6.0
Relevant Packages:
next: 15.2.0-canary.14 // Latest available version is detected (15.2.0-canary.14).
eslint-config-next: N/A
react: 19.0.0
react-dom: 19.0.0
typescript: 5.7.3
Next.js Config:
output: N/A
```
### Which area(s) are affected? (Select all that apply)
dynamicIO, Partial Prerendering (PPR)
### Which stage(s) are affected? (Select all that apply)
Vercel (Deployed), next start (local), next build (local)
### Additional context
The issue is being caused by an incorrect check in listenForDynamicRequest. The issue occurs after a rebuild of the server, this could be a local rebuild and restart of the server but also affects a deployment to Vercel without skew protection. Also note how the issue does not happen on the Not Prefetched link, it only affects prefetched pages | Partial Prerendering (PPR),dynamicIO | low | Minor |
2,796,115,021 | rust | new build_helper::compiletest module for safer invocations of compiletest by bootstrap | currently bootstrap just uses strings to represent various compiletest arguments, such as mode.
obviously it would be better to use enums for these, as they are much less vulnerable to typos.
`compiletest` already has enum definitions for these (in `common.rs`), and the two already have a shared dependency (`build_helper`), and `build_helper` already has access to `serde`.
i would be willing to work on this refactor, if people think it would be a good idea. | C-cleanup,T-bootstrap,A-compiletest | low | Minor |
2,796,117,237 | rust | ICE: polonius: `None` in region infer with unconditional panic | <!--
[31mICE[0m: Rustc ./a.rs '--edition=2024 -Zpolonius=next -ooutputfile -Zdump-mir-dir=dir' 'thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10: 'called `Option::unwrap()` on a `None` value'', 'thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10: 'called `Option::unwrap()` on a `None` value''
File: /tmp/im/a.rs
-->
not sure why this only panics in edition 2024 🤔
auto-reduced (treereduce-rust):
````rust
//@compile-flags: --edition=2024 -Zpolonius=next
fn main() {
&{ [1, 2, 3][4] };
}
````
original:
````rust
//@ build-pass
//@ ignore-pass (test emits codegen-time warnings and verifies that they are not errors)
#![warn(unconditional_panic)]
fn main() {
&{ [1, 2, 3][4] };
//~^ WARN operation will panic
}
````
Version information
````
rustc 1.86.0-nightly (bcd0683e5 2025-01-17)
binary: rustc
commit-hash: bcd0683e5dce1945b5d940714742e7502883bb5c
commit-date: 2025-01-17
host: x86_64-unknown-linux-gnu
release: 1.86.0-nightly
LLVM version: 19.1.7
````
Possibly related line of code:
https://github.com/rust-lang/rust/blob/bcd0683e5dce1945b5d940714742e7502883bb5c/compiler/rustc_borrowck/src/region_infer/mod.rs#L1882-L1894
Command:
`/home/matthias/.rustup/toolchains/master/bin/rustc --edition=2024 -Zpolonius=next`
<details><summary><strong>Program output</strong></summary>
<p>
```
thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10:
called `Option::unwrap()` on a `None` value
stack backtrace:
0: 0x7077524fb0ca - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::h245d3d13c9e117ef
1: 0x707752c12d26 - core::fmt::write::ha5c79450f6a60aba
2: 0x707753b95891 - std::io::Write::write_fmt::hc0bb623be5e87022
3: 0x7077524faf22 - std::sys::backtrace::BacktraceLock::print::h5f0375c539e6bbfe
4: 0x7077524fd3a2 - std::panicking::default_hook::{{closure}}::hfaae0dc7005f1f46
5: 0x7077524fd22a - std::panicking::default_hook::had3887552141764d
6: 0x70775165e33b - std[970739192a5a8910]::panicking::update_hook::<alloc[c190f769f0908da9]::boxed::Box<rustc_driver_impl[98a1c83b91f31d14]::install_ice_hook::{closure#1}>>::{closure#0}
7: 0x7077524fdee3 - std::panicking::rust_panic_with_hook::h0f6e30840497503a
8: 0x7077524fdba6 - std::panicking::begin_panic_handler::{{closure}}::h00d1f33c4abe282b
9: 0x7077524fb5a9 - std::sys::backtrace::__rust_end_short_backtrace::h7f80f131d37aefa4
10: 0x7077524fd89d - rust_begin_unwind
11: 0x70774f1a8530 - core::panicking::panic_fmt::h313dbd2ed26f8657
12: 0x70774fcc082c - core::panicking::panic::h6bed83e7eae20130
13: 0x70775067d5a9 - core::option::unwrap_failed::h5e1ae1d3b98bd9cb
14: 0x70775143be47 - <rustc_borrowck[4bd42f5d1352065e]::region_infer::RegionInferenceContext>::find_sub_region_live_at
15: 0x707750fd0fb4 - <rustc_borrowck[4bd42f5d1352065e]::MirBorrowckCtxt>::explain_why_borrow_contains_point
16: 0x707750e18b6b - <rustc_borrowck[4bd42f5d1352065e]::MirBorrowckCtxt>::report_borrowed_value_does_not_live_long_enough
17: 0x707753cf134f - rustc_borrowck[4bd42f5d1352065e]::do_mir_borrowck
18: 0x707753cc352f - rustc_query_impl[a1940b4b03f25947]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[a1940b4b03f25947]::query_impl::mir_borrowck::dynamic_query::{closure#2}::{closure#0}, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 8usize]>>
19: 0x707752f2260e - rustc_query_system[f87afe778d5ed02f]::query::plumbing::try_execute_query::<rustc_query_impl[a1940b4b03f25947]::DynamicConfig<rustc_data_structures[7c7a3305c98b4df4]::vec_cache::VecCache<rustc_span[e90766ed4fb19497]::def_id::LocalDefId, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 8usize]>, rustc_query_system[f87afe778d5ed02f]::dep_graph::graph::DepNodeIndex>, false, false, false>, rustc_query_impl[a1940b4b03f25947]::plumbing::QueryCtxt, false>
20: 0x707752f22113 - rustc_query_impl[a1940b4b03f25947]::query_impl::mir_borrowck::get_query_non_incr::__rust_end_short_backtrace
21: 0x707752f16fb0 - rustc_interface[f2cc204a1e18dc42]::passes::run_required_analyses
22: 0x707753b9109e - rustc_interface[f2cc204a1e18dc42]::passes::analysis
23: 0x707753b9106f - rustc_query_impl[a1940b4b03f25947]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[a1940b4b03f25947]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 0usize]>>
24: 0x707753be8355 - rustc_query_system[f87afe778d5ed02f]::query::plumbing::try_execute_query::<rustc_query_impl[a1940b4b03f25947]::DynamicConfig<rustc_query_system[f87afe778d5ed02f]::query::caches::SingleCache<rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 0usize]>>, false, false, false>, rustc_query_impl[a1940b4b03f25947]::plumbing::QueryCtxt, false>
25: 0x707753be808e - rustc_query_impl[a1940b4b03f25947]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace
26: 0x707753c5c2e9 - rustc_interface[f2cc204a1e18dc42]::passes::create_and_enter_global_ctxt::<core[31d5326fba56f7df]::option::Option<rustc_interface[f2cc204a1e18dc42]::queries::Linker>, rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}::{closure#2}>::{closure#2}::{closure#0}
27: 0x707753c4f4d6 - rustc_interface[f2cc204a1e18dc42]::interface::run_compiler::<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1}
28: 0x707753a92c07 - std[970739192a5a8910]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_with_globals<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_pool_with_globals<rustc_interface[f2cc204a1e18dc42]::interface::run_compiler<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>
29: 0x707753a928d9 - <<std[970739192a5a8910]::thread::Builder>::spawn_unchecked_<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_with_globals<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_pool_with_globals<rustc_interface[f2cc204a1e18dc42]::interface::run_compiler<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>::{closure#1} as core[31d5326fba56f7df]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0}
30: 0x707753a9206b - std::sys::pal::unix::thread::Thread::new::thread_start::h84cef28dceb72f89
31: 0x70774dea339d - <unknown>
32: 0x70774df2849c - <unknown>
33: 0x0 - <unknown>
error: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md
note: please make sure that you have updated to the latest nightly
note: rustc 1.86.0-nightly (bcd0683e5 2025-01-17) running on x86_64-unknown-linux-gnu
note: compiler flags: -Z polonius=next -Z dump-mir-dir=dir
query stack during panic:
#0 [mir_borrowck] borrow-checking `main`
#1 [analysis] running analysis passes on this crate
end of query stack
```
</p>
</details>
<!--
query stack:
#0 [mir_borrowck] borrow-checking `main`
#1 [analysis] running analysis passes on this crate
-->
| A-diagnostics,I-ICE,C-bug,NLL-polonius,T-types | low | Critical |
2,796,123,769 | rust | universally directly depend on serde_derive instead of accessing it through serde. | bootstrap does this in order to flatten its dependencies and improve build times, but since other crates (like `compiletest` and `rustdoc`) use the `derive` feature flag, i believe this means `serde` has to be built twice to use any of these programs.
wouldn't it be best to always use the same set of feature flags to serde? | C-cleanup,T-bootstrap | low | Minor |
2,796,174,444 | pytorch | `torch.onnx.export` (dynamo=False) fails with uninformative error when exporting `apply_rotary_pos_emb`/`repeat_interleave` | ### 🐛 Describe the bug
When attempting to export a HF transformers model that performs `apply_rotary_pos_emb` (definition below), an uninformative error is thrown:
<details>
<summary>See reproduction code</summary>
```py
import torch
def rotate_half(x):
"""Rotates half the hidden dims of the input."""
x1 = x[..., 0::2]
x2 = x[..., 1::2]
return torch.stack((-x2, x1), dim=-1).flatten(-2)
def apply_rotary_pos_emb(q, k, cos, sin, position_ids=None, unsqueeze_dim=1):
"""Applies Rotary Position Embedding to the query and key tensors.
Args:
q (`torch.Tensor`): The query tensor.
k (`torch.Tensor`): The key tensor.
cos (`torch.Tensor`): The cosine part of the rotary embedding.
sin (`torch.Tensor`): The sine part of the rotary embedding.
position_ids (`torch.Tensor`, *optional*):
Deprecated and unused.
unsqueeze_dim (`int`, *optional*, defaults to 1):
The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and
sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note
that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and
k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes
cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have
the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2.
Returns:
`tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding.
"""
cos = cos.unsqueeze(unsqueeze_dim)
sin = sin.unsqueeze(unsqueeze_dim)
# Interleave them instead of usual shape
cos = cos[..., : cos.shape[-1] // 2].repeat_interleave(2, dim=-1)
sin = sin[..., : sin.shape[-1] // 2].repeat_interleave(2, dim=-1)
# Keep half or full tensor for later concatenation
rotary_dim = cos.shape[-1]
q_rot, q_pass = q[..., :rotary_dim], q[..., rotary_dim:]
k_rot, k_pass = k[..., :rotary_dim], k[..., rotary_dim:]
# Apply rotary embeddings on the first half or full tensor
q_embed = (q_rot * cos) + (rotate_half(q_rot) * sin)
k_embed = (k_rot * cos) + (rotate_half(k_rot) * sin)
# Concatenate back to full shape
q_embed = torch.cat([q_embed, q_pass], dim=-1)
k_embed = torch.cat([k_embed, k_pass], dim=-1)
return q_embed, k_embed
class Model(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, q, k, cos, sin):
return apply_rotary_pos_emb(q, k, cos, sin)
model = Model()
torch.manual_seed(0)
q = torch.randn(2, 2, 40, 32)
k = torch.randn(2, 2, 40, 32)
cos = torch.randn(1, 40, 28)
sin = torch.randn(1, 40, 28)
output = model(q, k, cos, sin)
assert output[0].shape == q.shape
assert output[1].shape == k.shape
assert abs(output[0].mean().item() - -0.022467557340860367) < 1e-8
assert abs(output[1].mean().item() - 0.0071802930906414986) < 1e-8
torch.onnx.export(
model,
(q, k, cos, sin),
"model.onnx",
input_names=["q", "k", "cos", "sin"],
output_names=["q_embed", "k_embed"],
opset_version=18,
dynamo=False, # True works, False is broken
)
```
</details>
produces:
```
Traceback (most recent call last):
File "/workspaces/optimum/o/rot.py", line 74, in <module>
torch.onnx.export(
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/__init__.py", line 375, in export
export(
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 502, in export
_export(
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1564, in _export
graph, params_dict, torch_out = _model_to_graph(
^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1117, in _model_to_graph
graph = _optimize_graph(
^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 639, in _optimize_graph
graph = _C._jit_pass_onnx(graph, operator_export_type)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1836, in _run_symbolic_function
return symbolic_fn(graph_context, *inputs, **attrs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_helper.py", line 369, in wrapper
return fn(g, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_opset11.py", line 519, in cat
return opset9.cat(g, tensor_list, dim)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_helper.py", line 281, in wrapper
return fn(g, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_opset9.py", line 563, in cat
assert all(
AssertionError
```
The issue seems to be caused by the `repeat_interleave` call. However, the error message does not describe this.
NOTE: The model exports correctly with dynamo=True, and I understand that the previous exporter is not being worked on, so feel free to close this issue if needed. I raise it for these reasons:
- [Optimum](http://github.com/huggingface/optimum) still uses the previous exporter, and so this issue is present.
- I haven't been able to find this issue reported, so I'm adding it so I can link to it in my workaround (rewriting the `repeat_interleave` function).
- To possibly get a more informative error (since the current one doesn't indicate what the problem is)
### Versions
Collecting environment information...
PyTorch version: 2.5.1+cu124
Is debug build: False
CUDA used to build PyTorch: 12.4
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.6 LTS (x86_64)
GCC version: (Ubuntu 9.4.0-1ubuntu1~20.04.2) 9.4.0
Clang version: 10.0.0-4ubuntu1
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.12.1 (main, Sep 30 2024, 17:05:21) [GCC 9.4.0] (64-bit runtime)
Python platform: Linux-6.5.0-1025-azure-x86_64-with-glibc2.31
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
Address sizes: 48 bits physical, 48 bits virtual
CPU(s): 2
On-line CPU(s) list: 0,1
Thread(s) per core: 2
Core(s) per socket: 1
Socket(s): 1
NUMA node(s): 1
Vendor ID: AuthenticAMD
CPU family: 25
Model: 1
Model name: AMD EPYC 7763 64-Core Processor
Stepping: 1
CPU MHz: 3243.502
BogoMIPS: 4890.85
Virtualization: AMD-V
Hypervisor vendor: Microsoft
Virtualization type: full
L1d cache: 32 KiB
L1i cache: 32 KiB
L2 cache: 512 KiB
L3 cache: 32 MiB
NUMA node0 CPU(s): 0,1
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode
Vulnerability Spec store bypass: Vulnerable
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl tsc_reliable nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext invpcid_single vmmcall fsgsbase bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr rdpru arat npt nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold v_vmsave_vmload umip vaes vpclmulqdq rdpid fsrm
Versions of relevant libraries:
[pip3] numpy==2.0.2
[pip3] nvidia-cublas-cu12==12.4.5.8
[pip3] nvidia-cuda-cupti-cu12==12.4.127
[pip3] nvidia-cuda-nvrtc-cu12==12.4.127
[pip3] nvidia-cuda-runtime-cu12==12.4.127
[pip3] nvidia-cudnn-cu12==9.1.0.70
[pip3] nvidia-cufft-cu12==11.2.1.3
[pip3] nvidia-curand-cu12==10.3.5.147
[pip3] nvidia-cusolver-cu12==11.6.1.9
[pip3] nvidia-cusparse-cu12==12.3.1.170
[pip3] nvidia-nccl-cu12==2.21.5
[pip3] nvidia-nvjitlink-cu12==12.4.127
[pip3] nvidia-nvtx-cu12==12.4.127
[pip3] onnx==1.17.0
[pip3] onnx-graphsurgeon==0.5.2
[pip3] onnxconverter-common==1.14.0
[pip3] onnxruntime==1.20.1
[pip3] onnxscript==0.1.0.dev20241219
[pip3] onnxslim==0.1.42
[pip3] torch==2.5.1
[pip3] torchvision==0.20.1
[pip3] triton==3.1.0
[conda] No relevant packages | module: onnx,triaged,OSS contribution wanted | low | Critical |
2,796,188,005 | rust | ICE: `cannot convert to a region vid` | <!--
[31mICE[0m: Rustc ./a.rs '-Zincremental-verify-ich=yes -Cincremental=<dir> -Cdebuginfo=2 -Clink-dead-code=true -Zvalidate-mir --edition=2024' 'error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid', 'error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid'
File: /tmp/im2/a.rs
-->
snippet:
````rust
struct A<'a> {
x: Vec<A> = Vec::new(),
}
````
Version information
````
rustc 1.86.0-nightly (bcd0683e5 2025-01-17)
binary: rustc
commit-hash: bcd0683e5dce1945b5d940714742e7502883bb5c
commit-date: 2025-01-17
host: x86_64-unknown-linux-gnu
release: 1.86.0-nightly
LLVM version: 19.1.7
````
Possibly related line of code:
https://github.com/rust-lang/rust/blob/bcd0683e5dce1945b5d940714742e7502883bb5c/compiler/rustc_borrowck/src/universal_regions.rs#L882-L894
Command:
`/home/matthias/.rustup/toolchains/master/bin/rustc`
<details><summary><strong>Program output</strong></summary>
<p>
```
error[E0106]: missing lifetime specifier
--> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:2:12
|
2 | x: Vec<A> = Vec::new(),
| ^ expected named lifetime parameter
|
help: consider using the `'a` lifetime
|
2 | x: Vec<A<'a>> = Vec::new(),
| ++++
error[E0658]: default values on fields are experimental
--> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:2:14
|
2 | x: Vec<A> = Vec::new(),
| ^^^^^^^^^^^^^
|
= note: see issue #132162 <https://github.com/rust-lang/rust/issues/132162> for more information
= help: add `#![feature(default_field_values)]` to the crate attributes to enable
= note: this compiler was built on 2025-01-17; consider upgrading it if it is out of date
error[E0601]: `main` function not found in crate `mvce`
--> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:3:2
|
3 | }
| ^ consider adding a `main` function to `/tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs`
error[E0392]: lifetime parameter `'a` is never used
--> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:1:10
|
1 | struct A<'a> {
| ^^ unused lifetime parameter
|
= help: consider removing `'a`, referring to it in a field, or using a marker such as `PhantomData`
error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid
thread 'rustc' panicked at compiler/rustc_borrowck/src/universal_regions.rs:888:36:
Box<dyn Any>
stack backtrace:
0: 0x730ea98f502a - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::h4fb5b0de48364101
1: 0x730eaa012d66 - core::fmt::write::h6eb54c31028066ca
2: 0x730eaafa43d1 - std::io::Write::write_fmt::hd9e9dfb2c5e220cc
3: 0x730ea98f4e82 - std::sys::backtrace::BacktraceLock::print::ha2ddb07494fe7b3b
4: 0x730ea98f7302 - std::panicking::default_hook::{{closure}}::h4eed2bdbab3e3d9c
5: 0x730ea98f718a - std::panicking::default_hook::h520b49f6397a0732
6: 0x730ea8a58e1b - std[3244471bfd6eaa61]::panicking::update_hook::<alloc[8ba9d48c156baf43]::boxed::Box<rustc_driver_impl[df45097da6562dc6]::install_ice_hook::{closure#1}>>::{closure#0}
7: 0x730ea98f7e43 - std::panicking::rust_panic_with_hook::ha20255168d758ad1
8: 0x730ea8a94761 - std[3244471bfd6eaa61]::panicking::begin_panic::<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug>::{closure#0}
9: 0x730ea8a89666 - std[3244471bfd6eaa61]::sys::backtrace::__rust_end_short_backtrace::<std[3244471bfd6eaa61]::panicking::begin_panic<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug>::{closure#0}, !>
10: 0x730ea8a89423 - std[3244471bfd6eaa61]::panicking::begin_panic::<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug>
11: 0x730ea8a9e691 - <rustc_errors[af9a6eaf1dbd0244]::diagnostic::BugAbort as rustc_errors[af9a6eaf1dbd0244]::diagnostic::EmissionGuarantee>::emit_producing_guarantee
12: 0x730ea907bc53 - rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt::<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0}
13: 0x730ea90615da - rustc_middle[352907c3633efb2c]::ty::context::tls::with_opt::<rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0}, !>::{closure#0}
14: 0x730ea906146b - rustc_middle[352907c3633efb2c]::ty::context::tls::with_context_opt::<rustc_middle[352907c3633efb2c]::ty::context::tls::with_opt<rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0}, !>::{closure#0}, !>
15: 0x730ea6f596b0 - rustc_middle[352907c3633efb2c]::util::bug::bug_fmt
16: 0x730eab54beb3 - <rustc_borrowck[55cebfbcecacd79c]::universal_regions::UniversalRegionIndices>::to_region_vid.cold
17: 0x730eaa060745 - <rustc_middle[352907c3633efb2c]::ty::Ty as rustc_type_ir[ba8c709077c11182]::fold::TypeSuperFoldable<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>::try_super_fold_with::<rustc_type_ir[ba8c709077c11182]::fold::RegionFolder<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>
18: 0x730eaa05f068 - <rustc_middle[352907c3633efb2c]::ty::Ty as rustc_type_ir[ba8c709077c11182]::fold::TypeSuperFoldable<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>::try_super_fold_with::<rustc_type_ir[ba8c709077c11182]::fold::RegionFolder<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>
19: 0x730eab0c21f6 - rustc_borrowck[55cebfbcecacd79c]::do_mir_borrowck
20: 0x730eab0bfbef - rustc_query_impl[43ade5033af9534a]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[43ade5033af9534a]::query_impl::mir_borrowck::dynamic_query::{closure#2}::{closure#0}, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 8usize]>>
21: 0x730eaa33c04e - rustc_query_system[eb70713c06242342]::query::plumbing::try_execute_query::<rustc_query_impl[43ade5033af9534a]::DynamicConfig<rustc_data_structures[dc2c66faf0aa2c5f]::vec_cache::VecCache<rustc_span[2bc6ccfbd90abceb]::def_id::LocalDefId, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 8usize]>, rustc_query_system[eb70713c06242342]::dep_graph::graph::DepNodeIndex>, false, false, false>, rustc_query_impl[43ade5033af9534a]::plumbing::QueryCtxt, false>
22: 0x730eaa33bb53 - rustc_query_impl[43ade5033af9534a]::query_impl::mir_borrowck::get_query_non_incr::__rust_end_short_backtrace
23: 0x730eaa3309f0 - rustc_interface[508537c9f518ed83]::passes::run_required_analyses
24: 0x730eaaf9fade - rustc_interface[508537c9f518ed83]::passes::analysis
25: 0x730eaaf9faaf - rustc_query_impl[43ade5033af9534a]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[43ade5033af9534a]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 0usize]>>
26: 0x730eaaf5a9d5 - rustc_query_system[eb70713c06242342]::query::plumbing::try_execute_query::<rustc_query_impl[43ade5033af9534a]::DynamicConfig<rustc_query_system[eb70713c06242342]::query::caches::SingleCache<rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 0usize]>>, false, false, false>, rustc_query_impl[43ade5033af9534a]::plumbing::QueryCtxt, false>
27: 0x730eaaf5a70e - rustc_query_impl[43ade5033af9534a]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace
28: 0x730eab0391e9 - rustc_interface[508537c9f518ed83]::passes::create_and_enter_global_ctxt::<core[24a92a8ec63ee5f4]::option::Option<rustc_interface[508537c9f518ed83]::queries::Linker>, rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}::{closure#2}>::{closure#2}::{closure#0}
29: 0x730eab02c416 - rustc_interface[508537c9f518ed83]::interface::run_compiler::<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1}
30: 0x730eaae73a07 - std[3244471bfd6eaa61]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[508537c9f518ed83]::util::run_in_thread_with_globals<rustc_interface[508537c9f518ed83]::util::run_in_thread_pool_with_globals<rustc_interface[508537c9f518ed83]::interface::run_compiler<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>
31: 0x730eaae736d9 - <<std[3244471bfd6eaa61]::thread::Builder>::spawn_unchecked_<rustc_interface[508537c9f518ed83]::util::run_in_thread_with_globals<rustc_interface[508537c9f518ed83]::util::run_in_thread_pool_with_globals<rustc_interface[508537c9f518ed83]::interface::run_compiler<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>::{closure#1} as core[24a92a8ec63ee5f4]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0}
32: 0x730eaae72e6f - std::sys::pal::unix::thread::Thread::new::thread_start::h5d259864ca2af1e4
33: 0x730ea52a339d - <unknown>
34: 0x730ea532849c - <unknown>
35: 0x0 - <unknown>
note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md
note: please make sure that you have updated to the latest nightly
note: please attach the file at `/tmp/im2/rustc-ice-2025-01-17T19_45_25-2933219.txt` to your bug report
query stack during panic:
#0 [mir_borrowck] borrow-checking `A::x::{constant#0}`
#1 [analysis] running analysis passes on this crate
end of query stack
error: aborting due to 5 previous errors
Some errors have detailed explanations: E0106, E0392, E0601, E0658.
For more information about an error, try `rustc --explain E0106`.
```
</p>
</details>
<!--
query stack:
note: compiler flags: -Z incremental-verify-ich=yes -C incremental=[REDACTED] -C debuginfo=2 -C link-dead-code=true -Z validate-mir
#0 [mir_borrowck] borrow-checking `A::x::{constant#0}`
#1 [analysis] running analysis passes on this crate
-->
| I-ICE,T-compiler,C-bug | low | Critical |
2,796,198,136 | deno | Error when using TS Decorators | Version: Deno 2.1.6
Error when using argument decorators
 | needs investigation | low | Critical |
2,796,212,730 | rust | Change in leaf function causes unrelated compilation error in root function | ## Description
I encountered an issue where changing the leaf function causes a compilation error in the root function, but the change doesn't affect types or auto traits of the function.
Commenting out `router()` makes code to compile.
## Minimal Example
```rust
use axum::routing::{post, Router};
use futures_util::StreamExt;
use std::collections::HashMap;
use std::sync::Arc;
struct Transaction {
f1: (),
f2: (),
}
pub(crate) struct SplitAddress(());
async fn handler_inner(ctx: Context) {
let transactions = vec![];
// This causes the lifetime issue:
let iter = transactions
.iter()
.map(|x: &Transaction| (x.f1, SplitAddress(x.f2)));
// This works:
// let iter: Vec<_> = transactions
// .iter()
// .map(|x: &Transaction| (x.f1, SplitAddress(x.f2)))
// .collect();
ctx.0.streaming_work(iter.into_iter()).await;
}
async fn handler() -> Result<String, String> {
let ctx = Context(Arc::new(Service()));
handler_inner(ctx).await;
Ok("ok".to_string())
}
// or comment out router
pub async fn router() -> Router {
Router::new().route("/bla", post(handler))
}
#[derive(Clone)]
pub struct Context(Arc<Service>);
struct Service();
impl Service {
pub(crate) async fn streaming_work(
&self,
data: impl Iterator<Item = ((), SplitAddress)>,
) -> HashMap<(), Vec<String>> {
futures_util::stream::iter(data)
.map(|_| async move { todo!() })
.buffer_unordered(100)
.filter_map(|x| async move { x })
.collect()
.await
}
}
```
[repo](https://github.com/0xdeafbeef/rustc-issue)
## Error Message
```
error: implementation of `FnOnce` is not general enough
--> src/lib.rs:39:33
|
39 | Router::new().route("/bla", post(handler))
| ^^^^^^^^^^^^^ implementation of `FnOnce` is not general enough
|
= note: closure with signature `fn(&'0 Transaction) -> ((), SplitAddress)` must implement `FnOnce<(&'1 Transaction,)>`, for any two lifetimes `'0` and `'1`...
= note: ...but it actually implements `FnOnce<(&Transaction,)>`
error: could not compile `rust-issue` (lib) due to 1 previous error
``` | A-lifetimes,A-trait-system,needs-triage | low | Critical |
2,796,227,000 | pytorch | torch._C._IncludeDispatchKeyGuard is very broken? | ```py
import torch
print(torch._C._meta_in_tls_dispatch_include())
with torch._C._IncludeDispatchKeyGuard(torch.DispatchKey.Meta):
print(torch._C._meta_in_tls_dispatch_include())
print(torch._C._meta_in_tls_dispatch_include())
```
prints False, True, True, which is completely bogus | triaged,module: dispatch | low | Critical |
2,796,232,607 | godot | Camera3D preview may show temporary visual glitch when selected | ### Tested versions
- Godot v4.4.beta1
- related https://github.com/godotengine/godot/pull/90778
### System information
Windows 11 (build 22631) - Single-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2070 (NVIDIA; 32.0.15.6614) - Intel(R) Core(TM) i7-10750H CPU @ 2.60GHz (12 threads)
### Issue description
When there are multiple cameras and you switch between them, there may be a short visual glitch in the camera preview.
It doesn't really need multiple cameras to happen, but it happens less often.
https://github.com/user-attachments/assets/92811fc9-99c5-4873-92e3-15a639ddc09d
Maybe related: the preview updates late in some cases, see https://github.com/godotengine/godot/pull/90778#issuecomment-2395606441
The preview should not show anything until it is ready.
### Steps to reproduce
1. Add 2 Camera3Ds to a scene.
2. Repeatedly switch between them while watching the preview.
3. You may see visual glitches.
### Minimal reproduction project (MRP)
Not really needed but here:
[camera_preview_multiple_2025-01-17_15-18-50.zip](https://github.com/user-attachments/files/18459220/camera_preview_multiple_2025-01-17_15-18-50.zip) | bug,topic:editor | low | Minor |
2,796,236,006 | vscode | Remote SSH freezing when loading extensions. Enabling remote.downloadExtensionsLocally did not help |
Type: <b>Bug</b>
In the server output I'm seeing:
2025-01-17 14:08:27.683 [error] [network] #1: https://main.vscode-cdn.net/extensions/marketplace.json - error GET read ETIMEDOUT
Yes I know remote SSH can be a problem if the remote server is behind a firewall. However I was hoping that the Extension Host could still work even in this case with a setting to download extensions locally.
Unfortunately this remote file is not working. Can you check to see if this was recently introduced?
VS Code version: Code 1.96.3 (91fbdddc47bc9c09064bf7acf133d22631cbf083, 2025-01-09T18:14:09.060Z)
OS version: Windows_NT x64 10.0.22631
Modes:
Remote OS version: Linux x64 5.14.0-437.el9.x86_64
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) Ultra 7 165U (14 x 2688)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|undefined|
|Memory (System)|31.46GB (5.65GB free)|
|Process Argv|--crash-reporter-id 7e9e6a5c-e81d-4f36-9073-13611d70a8a4|
|Screen Reader|no|
|VM|0%|
|Item|Value|
|---|---|
|Remote|SSH: jumphost|
|OS|Linux x64 5.14.0-437.el9.x86_64|
|CPUs|Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz (2 x 0)|
|Memory (System)|3.82GB (1.96GB free)|
|VM|100%|
</details><details><summary>Extensions (5)</summary>
Extension|Author (truncated)|Version
---|---|---
catppuccin-vsc|Cat|3.16.0
remote-ssh|ms-|0.116.1
remote-ssh-edit|ms-|0.87.0
remote-explorer|ms-|0.4.3
ufuk-bakan-color-themes|Ufu|2.0.1
(3 theme extensions excluded)
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vspor879:30202332
vspor708:30202333
vspor363:30204092
vscod805:30301674
binariesv615:30325510
vsaa593:30376534
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
2i9eh265:30646982
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
h48ei257:31000450
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupyter:31046869
nativerepl2:31139839
pythonrstrctxt:31112756
nativeloc2:31192216
cf971741:31144450
iacca1:31171482
notype1:31157159
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
390bf810:31215807
6074i472:31201624
dwoutputs:31217127
```
</details>
<!-- generated by issue reporter --> | triage-needed,stale | low | Critical |
2,796,242,535 | material-ui | top-level-imports v5 changing color imports | ### Steps to reproduce
Steps:
1. Import grey from @mui/material/colors
`import { grey } from @mui/material/colors;`
2. npx @mui/codemod@latest v5.0.0/top-level-imports <path>
3. The import will change to `import { grey } from '@mui/material';`
### Current behavior
Steps:
1. Import grey from @mui/material/colors
`import { grey } from @mui/material/colors;`
2. npx @mui/codemod@latest v5.0.0/top-level-imports <path>
3. The import will change to `import { grey } from '@mui/material';`
### Expected behavior
`npx @mui/codemod@latest v5.0.0/top-level-imports <path>` should not change imports to those that are not valid
### Context
_No response_
### Your environment
<details>
<summary><code>npx @mui/envinfo</code></summary>
```
System:
OS: macOS 15.1.1
Binaries:
Node: 21.2.0 - ~/node_modules/.bin/node
npm: 10.7.0 - ~/.nvm/versions/node/v18.20.3/bin/npm
pnpm: Not Found
Browsers:
Chrome: 132.0.6834.83
Edge: 131.0.2903.147
Safari: 18.1.1
npmPackages:
@emotion/react: ^11.10.4 => 11.14.0
@emotion/styled: ^11.10.4 => 11.14.0
@mui/base: 5.0.0-beta.66
@mui/core-downloads-tracker: 5.16.11
@mui/icons-material: ^5.8.4 => 5.16.11
@mui/lab: ^5.0.0-alpha.95 => 5.0.0-alpha.174
@mui/material: ^5.10.1 => 5.16.11
@mui/material-nextjs: ^5.15.6 => 5.16.8
@mui/private-theming: 5.16.8
@mui/styled-engine: 5.16.8
@mui/styles: ^5.9.3 => 5.16.11
@mui/system: 5.16.8
@mui/types: 7.2.19
@mui/utils: 5.16.8
@mui/x-data-grid: 6.20.4
@mui/x-data-grid-pro: ^6.0.4 => 6.20.4
@mui/x-date-pickers: 6.20.2
@mui/x-date-pickers-pro: ^6.0.0 => 6.20.2
@mui/x-license-pro: 6.10.2
@types/react: ^18.0.28 => 18.3.16
react: ^18.2.0 => 18.3.1
react-dom: ^18.2.0 => 18.3.1
typescript: ^4.9.5 => 4.9.5
```
</details>
**Search keywords**: codemod, top-level-import, color imports | bug 🐛,waiting for 👍,package: codemod,ready to take,v5.x | low | Minor |
2,796,256,091 | tauri | [bug] `cargo tauri dev` does not kill the application before recompiling Rust/config changes on Windows | ### Describe the bug
I used `cargo install [email protected]` with toolchain 1.84 and started migrating a v1 project to v2. Everything seems to be working, except when I run the dev server with `cargo tauri dev` with the file watcher enabled, it no longer kills the child process before trying to rerun `cargo build` when I change the Rust code or the Tauri config. I think it might be due to changing the child process runner to use `cargo run`, and that it's killing the `cargo.exe` process but not the `appname.exe` process spawned by that.
### Reproduction
1. Create a Tauri React app.
2. Install the `tauri-cli` binary locally instead of using the version downloaded by `yarn`.
3. Run `cargo tauri dev`.
4. Touch/save a change to `src-tauri/tauri.conf.json`.
### Expected behavior
Along with killing the web dev server, and like on Unix-flavors, the app should be killed before rerunning `cargo build`. Otherwise, the file will still be in use and the dev server just errors out.
### Full `tauri info` output
```text
[⚠] Environment
- OS: Windows 10.0.26100 x86_64 (X64)
✔ WebView2: 131.0.2903.146
✔ MSVC: Visual Studio Enterprise 2022
✔ rustc: 1.84.0 [redacted]
✔ cargo: 1.84.0
⚠ rustup: not installed!
If you have rust installed some other way, we recommend uninstalling it
then use rustup instead. Visit https://rustup.rs/
⚠ Rust toolchain: couldn't be detected!
Maybe you don't have rustup installed? if so, Visit https://rustup.rs/
- node: 22.5.1
- yarn: 1.22.19
- npm: 10.8.2
[-] Packages
- tauri 🦀: 2.2.2
- tauri-build 🦀: 2.0.5
- wry 🦀: 0.48.1
- tao 🦀: 0.31.1
- tauri-cli 🦀: 2.2.4
- @tauri-apps/api : 2.2.0
- @tauri-apps/cli : 2.2.4
[-] Plugins
- tauri-plugin-fs 🦀: 2.2.0
- @tauri-apps/plugin-fs : 2.2.0
[-] App
- build-type: bundle
- CSP: unset
- frontendDist: ../dist
- devUrl: http://localhost:1420/
```
### Stack trace
```text
```
### Additional context
We periodically rebuild the Rust toolchain internally, which is why there's no default `rustup` and why I redacted some of the `rustc` fingerprint in the `tauri info` output. It's also why I always `cargo install tauri-cli` locally and run everything with `cargo tauri` rather than `yarn tauri`. | type: bug,status: needs triage | low | Critical |
2,796,258,184 | PowerToys | Task Scheduler fails to execute autorun | ### Microsoft PowerToys version
0.87.1
### Installation method
PowerToys auto-update
### Running as admin
Yes
### Area(s) with issue?
General
### Steps to reproduce
This is one of the "Autorun not working at startup" issues. The issue has only shown up within the past month after using PowerToys the last 4 years. I have been perusing the previous threads without much success, so I am sharing what I have attempted so far.
- Toggling autorun on/off in the general settings UI.
- Tweaking Task Scheduler's startup condition (e.g., launching at startup regardless of log on behaviour)
Now Task Scheduler is being odd here. It claims to have successfully executed the scheduled task, but the Last Run Result displays the 0x41303 string (Not Running). Running the task manually does start PowerToys.


### ✔️ Expected Behavior
PowerToys launching on startup
### ❌ Actual Behavior
PowerToys not launching on startup
### Other Software
_No response_ | Issue-Bug,Needs-Triage | low | Minor |
2,796,316,504 | ui | [bug]: CLI not respecting settings in components.json | ### Describe the bug
```json
{
"$schema": "https://ui.shadcn.com/schema.json",
"style": "new-york",
"rsc": false,
"tsx": true,
"tailwind": {
"config": "tailwind.config.ts",
"css": "app/app.css",
"baseColor": "zinc",
"cssVariables": true,
"prefix": "",
"tsx": true
},
"aliases": {
"components": "~/app/components",
"utils": "~/app/utilities",
"ui": "~/app/components/ui",
"lib": "~/app/utilities",
"hooks": "~/app/hooks"
},
"iconLibrary": "lucide"
}
```
I have this for my components.json and whenever I install a new component the import for `cn` is always `import { cn } from "~/lib/utils"`
### Affected component/components
CLI
### How to reproduce
1. Install shadcn
2. Modify `components.json` with a new `utils` alias
3. Install any component (i.e.: `pnpm dlx shadcn@latest add checkbox`)
4. Open the checkbox and see that the config wasn't respected
### Codesandbox/StackBlitz link
_No response_
### Logs
```bash
```
### System Info
```bash
MacOS, Vite, React Router 7
```
### Before submitting
- [x] I've made research efforts and searched the documentation
- [x] I've searched for existing issues | bug | low | Critical |
2,796,341,555 | PowerToys | [Github] PRs missing build CI checks | ### Repository Branch
main
### ❌ Actual Behavior
The build ci checks aren't running anymore after updating PRs against current main. Why?
xref: #33703

### ✔️ Expected Behavior
Build CI checks are executed after pushing changes. | Issue-Bug,Priority-0,Area-Build,Needs-Team-Response,Status-Reproducible | low | Minor |
2,796,343,916 | langchain | docs: Replace `initialize_agent` use with langgraph.prebuilt.create_react_agent | ### Privileged issue
- [x] I am a LangChain maintainer, or was asked directly by a LangChain maintainer to create an issue here.
### Issue Content
Currently there's a bunch of tools and other integrations that use `initialize_agent` in their docs page (search langchain/docs for "initialize_agent")
This function has been deprecated since 0.1 and should be replaced with basic langgraph usage as a demo
Deprecated function: https://python.langchain.com/api_reference/langchain/agents/langchain.agents.initialize.initialize_agent.html
Example page using create_react_agent in an example: https://python.langchain.com/docs/integrations/tools/gmail/#use-within-an-agent
This issue doesn't need to be resolved by a single PR and can be tackled incrementally! Just tag this issue for tracking purposes :) | documentation,help wanted,good first issue,🤖:docs | low | Minor |
2,796,350,085 | rust | Tracking issue for release notes of #135656: Add `-C hint-mostly-unused` to tell rustc that most of a crate will go unused |
This issue tracks the release notes text for #135656.
### Steps
- [ ] Proposed text is drafted by PR author (or team) making the noteworthy change.
- [ ] Issue is nominated for release team review of clarity for wider audience.
- [ ] Release team includes text in release notes/blog posts.
### Release notes text
The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing).
````markdown
# Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...)
- [Add `-C hint-mostly-unused` to tell rustc that most of a crate will go unused](https://github.com/rust-lang/rust/pull/135656)
````
> [!TIP]
> Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use.
> The category will be de-duplicated with all the other ones by the release team.
>
> *More than one section can be included if needed.*
### Release blog section
If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section.
*Otherwise leave it empty.*
````markdown
````
cc @joshtriplett, @fee1-dead -- origin issue/PR authors and assignees for starting to draft text
| T-compiler,relnotes,relnotes-tracking-issue | low | Minor |
2,796,353,567 | langchain | Update default values of minimax_api_host and model in MiniMaxChat __init__ | ### Checked other resources
- [x] I added a very descriptive title to this issue.
- [x] I searched the LangChain documentation with the integrated search.
- [x] I used the GitHub search to find a similar question and didn't find it.
- [x] I am sure that this is a bug in LangChain rather than my code.
- [x] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package).
### Example Code
The following code:
```python
from langchain_community.chat_models import MiniMaxChat
api_key = "YOUR_API_KEY"
chat = MiniMaxChat(api_key=api_key)
messages = [
("human", "Hello, how are you?"),
]
resp = chat.invoke(messages)
print(resp)
```
### Error Message and Stack Trace (if applicable)
```
Traceback (most recent call last):
File "/home/work/langchain/bug_report.py", line 13, in <module>
resp = chat.invoke(messages)
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 286, in invoke
self.generate_prompt(
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 786, in generate_prompt
return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs)
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 643, in generate
raise e
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 633, in generate
self._generate_with_cache(
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 851, in _generate_with_cache
result = self._generate(
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_community/chat_models/minimax.py", line 515, in _generate
return self._create_chat_result(response.json())
File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_community/chat_models/minimax.py", line 429, in _create_chat_result
for res in response["choices"]:
KeyError: 'choices'
```
### Description
* I am trying to run `MiniMaxChat`using default settings
* The underlying problem is that the default values for parameters do not work
- `minimax_api_host` is `https://api.minimax.chat/v1/text/chatcompletion_v2`, whereas it should be `https://api.minimaxi.chat/v1/text/chatcompletion_v2` (note the `i` -> `minimax` -> `minimaxi`)
- `model` is `abab6.5-chat`, whereas it should be `abab6.5s-chat` (note the `s`)
- Consequently, the received response is not a succesful one, so the `choices` key is not present in it
* Using overriden parameters `MiniMaxChat(api_key=api_key, model="abab6.5s-chat", minimax_api_host="https://api.minimaxi.chat/v1/text/chatcompletion_v2")` yields successful result
* Additionally there is a problem with the code itself at the point of the raised exception. It's difficult to comprehend for a user why the `KeyError` was raised - but that is for another issue probably.
- In my opinion, it would be better to fail early with an informative error message - invalid API key, insufficient balance, invalid model name etc.
### System Info
`python -m langchain_core.sys_info` yields:
```
System Information
------------------
> OS: Linux
> OS Version: #52~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Mon Dec 9 15:00:52 UTC 2
> Python Version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0]
Package Information
-------------------
> langchain_core: 0.3.30
> langchain: 0.3.14
> langchain_community: 0.3.14
> langsmith: 0.2.11
> langchain_minimax: Installed. No version info available.
> langchain_openai: 0.3.0
> langchain_text_splitters: 0.3.5
Optional packages not installed
-------------------------------
> langserve
Other Dependencies
------------------
> aiohttp: 3.11.11
> async-timeout: 4.0.3
> dataclasses-json: 0.6.7
> httpx: 0.28.1
> httpx-sse: 0.4.0
> jsonpatch: 1.33
> langsmith-pyo3: Installed. No version info available.
> numpy: 1.26.4
> openai: 1.59.8
> orjson: 3.10.14
> packaging: 24.2
> pydantic: 2.10.5
> pydantic-settings: 2.7.1
> PyYAML: 6.0.2
> requests: 2.32.3
> requests-toolbelt: 1.0.0
> SQLAlchemy: 2.0.37
> tenacity: 9.0.0
> tiktoken: 0.8.0
> typing-extensions: 4.12.2
> zstandard: Installed. No version info available.
``` | 🤖:bug | low | Critical |
2,796,356,520 | godot | [Godot 4.4 Beta1] Big TileMapLayer's tiles disappear and flicker when used inside a CanvasGroup and zooming out | ### Tested versions
Reproducible in: 4.4 Beta1, 4.4 dev7.
Not reproducible in: 4.1.4 stable.
### System information
Godot v4.4.beta1 - Windows 10 (build 19045) - Multi-window, 1 monitor - Vulkan (Forward+) - integrated AMD Radeon(TM) Vega 8 Graphics (Advanced Micro Devices, Inc.; 27.20.1028.1) - AMD Ryzen 5 3500U with Radeon Vega Mobile Gfx (8 threads)
### Issue description
I was switching a project from Godot 4.1.4 to 4.4 when I noticed the minimap I made for it disappeared, it was also flickering in and out every time the character moved.
The setup for it was really simple, so i made a test project to confirm if it also happened from a blank project and it did.
(I made a shader to make this easier to view in the reproduction project and works without it, the TileMapLayer outside the CanvasGroup has black tiles and the one inside the CanvasGroup has white tiles)

Zooming out

This works both in the editor and when running the project.
As to why I am using a CanvasGroup with a TileMapLayer this big: I made a shader in 4.1.4 using the texture CanvasGroup gives, it allowed me to make effects I really liked, but I can't do the same effects without it.
Changing the fit margin/clear margin/mipmaps in the CanvasGroup doesn't seem to do anything even when set to ridiculous values.
### Steps to reproduce
-Make a scene with a CanvasGroup.
-Add a TileMapLayer as a child of the CanvasGroup.
-Add a TileSet to the TileMapLayer and put any texture to it to use as tiles.
-Use the Rect Tool to fill a very large area of the world with tiles.
-Zoom the camera out.
### Minimal reproduction project (MRP)
[canvasgroupbigtilemap.zip](https://github.com/user-attachments/files/18460023/canvasgroupbigtilemap.zip) | bug,topic:editor,needs testing,topic:2d | low | Minor |
2,796,381,668 | go | cmd/go: `all` pattern in per package flags is inaccurate in module mode | `all` has a slightly different meaning in module mode than it does in `GOPATH` mode. In `GOPATH` mode, it's pretty much everything:
> In GOPATH mode, "all" expands to all packages found in all the GOPATH trees.
But in module mode, "all"
> expands to the set of "packages in the main module (or workspace modules) and their dependencies, including dependencies needed by tests of any of those"
according to `go help packages`.
We don't seem to handle this correctly when processing "all" in `-coverpkg` and the per package flags `-asmflags`, `-gccgoflags`, `-gcflags`, and `-ldflags`. The cmd/go/internal/load.MatchPackage function just matches every package for all.
| NeedsInvestigation,GoCommand | low | Minor |
2,796,383,244 | pytorch | Obey sm_carveout (limit on number of SMs) in inductor persistent kernel | ### 🚀 The feature, motivation and pitch
See https://github.com/pytorch/pytorch/pull/144974#issuecomment-2599011250
cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @aakhundov @eellison @lw
### Alternatives
_No response_
### Additional context
_No response_ | triaged,oncall: pt2,module: inductor | low | Minor |
2,796,397,162 | rust | E0106 "missing lifetime specifier" should be omitted when the signature is incorrect in an `impl` (E0186) | ### Code
```Rust
trait Foo {
fn foo(&self) -> &str;
}
struct Example;
impl Foo for Example {
fn foo() -> &str {
"example"
}
}
```
### Current output
```Shell
error[E0106]: missing lifetime specifier
--> src/lib.rs:7:17
|
7 | fn foo() -> &str {
| ^ expected named lifetime parameter
|
= help: this function's return type contains a borrowed value, but there is no value for it to be borrowed from
help: consider using the `'static` lifetime, but this is uncommon unless you're returning a borrowed value from a `const` or a `static`
|
7 | fn foo() -> &'static str {
| +++++++
help: instead, you are more likely to want to return an owned value
|
7 | fn foo() -> String {
| ~~~~~~
error[E0186]: method `foo` has a `&self` declaration in the trait, but not in the impl
--> src/lib.rs:7:5
|
2 | fn foo(&self) -> &str;
| ---------------------- `&self` used in trait
...
7 | fn foo() -> &str {
| ^^^^^^^^^^^^^^^^ expected `&self` in impl
```
### Desired output
```Shell
error[E0186]: method `foo` has a `&self` declaration in the trait, but not in the impl
--> src/lib.rs:7:5
|
2 | fn foo(&self) -> &str;
| ---------------------- `&self` used in trait
...
7 | fn foo() -> &str {
| ^^^^^^^^^^^^^^^^ expected `&self` in impl
```
### Rationale and extra context
When the function is a trait associated function, the signature must match (or refine) the trait’s, so E0106 and any other errors that would propose changes to the signature are misleading or at least not nearly as relevant as E0186 (or E0050 or E0053 or any other such mismatch). Therefore, they should be hidden.
This is particularly important for situations involving a missing `self` parameter, because (in my personal experience) it’s easy to forget to write `self` in the parameter list.
### Rust Version
```Shell
1.84.0
```
| A-diagnostics,T-compiler | low | Critical |
2,796,412,976 | react-native | Zooming in ScrollView causes incorrect offset of content | ### Description
When using the zoom and pinchGestureEnabled on a ScrollView, the content shifts to the side after re-rendering the component when zoomed in.
### Steps to reproduce
1) Toggle the ScrollView on
2) Zoom in
3) Toggle the ScrollView off
4) Toggle the ScrollView on
### React Native Version
0.76.6
### Affected Platforms
Runtime - iOS
### Output of `npx react-native info`
```text
System:
OS: Windows 11 10.0.22631
CPU: (12) x64 Intel(R) Core(TM) i7-10750H CPU @ 2.60GHz
Memory: 3.26 GB / 15.78 GB
Binaries:
Node:
version: 22.8.0
path: C:\Program Files\nodejs\node.EXE
Yarn:
version: 1.22.22
path: C:\Program Files\nodejs\yarn.CMD
npm:
version: 10.8.2
path: C:\Program Files\nodejs\npm.CMD
Watchman: Not Found
SDKs:
Android SDK: Not Found
Windows SDK: Not Found
IDEs:
Android Studio: Version 2021.1.0.0 AI-211.7628.21.2111.8193401
Visual Studio:
- 17.3.32804.467 (Visual Studio Community 2022)
Languages:
Java:
version: 17.0.8
path: /c/Program Files/Common Files/Oracle/Java/javapath/javac
Ruby: Not Found
npmPackages:
"@react-native-community/cli":
installed: 15.1.3
wanted: ^15.1.3
react:
installed: 18.3.1
wanted: 18.3.1
react-native:
installed: 0.76.6
wanted: 0.76.6
react-native-windows: Not Found
npmGlobalPackages:
"*react-native*": Not Found
Android:
hermesEnabled: Not found
newArchEnabled: Not found
iOS:
hermesEnabled: Not found
newArchEnabled: Not found
```
### Stacktrace or Logs
```text
Visual bug
```
### Reproducer
https://snack.expo.dev/@bram-dc/scrollview-zoom-offset
### Screenshots and Videos
https://github.com/user-attachments/assets/7eaed09f-1b25-4cb1-bb7d-379bf86b29d4 | Issue: Author Provided Repro,Component: ScrollView,Needs: Version Info | low | Critical |
2,796,457,228 | go | x/build/cmd/relui: add support for making release announcement posts on Bluesky | During Go releases, relui takes on posting on 2 social sites using their APIs (#47403 and #65046). By now we have an account on one more social site, Bluesky, where we've also started to post links to Go release announcements (for example, https://bsky.app/profile/golang.org/post/3lfvcgeox7c2c).
This is the tracking issue for automating Go release announcement posts on [Bluesky](https://docs.bsky.app/).
CC @golang/release, @mknyszek, @fflewddur. | Builders,NeedsInvestigation,Automation | low | Minor |
2,796,462,469 | flutter | [flutter_test] gestures can leak from test to test | Steps to reproduce:
* Check out this revision of Flutter: https://github.com/flutter/flutter/commit/bb2d34126cc8161dbe4a1bf23c925e48b732f670
* `cd packages/flutter`
* `flutter test --reporter=expanded --test-randomize-ordering-seed=20250118 --fatal-warnings --track-widget-creation test/material/reorderable_list_test.dart`
* Observe the test passing.
* Now change the seed: `flutter test --reporter=expanded --test-randomize-ordering-seed=20250117 --fatal-warnings --track-widget-creation test/material/reorderable_list_test.dart`
* Observe the test failing.
The reason it's failing is because a test in `reorderable_list_test.dart` created a `TestGesture` but never finalized it by cancelling it or by calling `drag.up()`. Because tests that are sensitive to dangling gestures could be anywhere, this issue did not immediately reproduce until, by chance, we rolled a seed that sorts tests such that the sensitive test was run immediately after the test that leaves the gesture hanging (which took about 1 month). This was fixed in https://github.com/flutter/flutter/pull/161836. However, the issue should have been caught immediately when `reorderable_list_test.dart` was added for the first time. `flutter_test` could validate that when a test ends, there are not dangling gestures and `VelocityTrackers` left.
| a: tests,framework,P1,team-framework,triaged-framework | medium | Minor |
2,796,500,511 | vscode | Line normalization in `Terminal.sendText()` (IPython) | <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
<details>
<summary>Show System Information</summary>
- VSCode: Version: 1.96.4 (user setup)
Commit: cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba
Date: 2025-01-16T00:16:19.038Z
Electron: 32.2.6
ElectronBuildId: 10629634
Chromium: 128.0.6613.186
Node.js: 20.18.1
V8: 12.8.374.38-electron.0
OS: Windows_NT x64 10.0.19045
- I'm using Windows 10, Version 22H2 with WSL2 (Ubuntu 22.04)
</details>
---
### 🎈 `Terminal.sendText()`
The `Terminal.sendText()` function currently normalizes line endings as follows:
https://github.com/microsoft/vscode/blob/0d57cf5304672f3cd0c2776aaec1e5dedfead053/src/vs/workbench/contrib/terminal/browser/terminalInstance.ts#L1271-L1275
With regards to cross-platform operability, I am wondering **why we don't append any `\n` here and instead just a single `\r`**? E.g. the default line separation on Windows is `\r\n` and not `\r`.
Git Blame: The mentioned code was first proposed in [this comment](https://github.com/microsoft/vscode/issues/22887#issuecomment-296505048) and then introduced via [this commit](https://github.com/microsoft/vscode/commit/d01c71579d6690e7a7c8311f8477275d8b9b0203). During this, the `os.EOL` was replaced by `\r` and I don't see the reason to do so.
---
### 🎈 Possible symptoms
I suspect this normalization to have caused [this problem](https://github.com/ipython/ipython/issues/13054#issuecomment-2599055070) for us in the role of VSCode extension authors. For that, assume that the extension user has opened a VSCode-integrated terminal and started an IPython shell inside it (IPython: interactive Python shell, feels like a REPL for Python). Assume further that the user has defined a Python method `def hello_world()` inside that IPython shell.
In our extension, we send the string `hello_world()` to that terminal and expect that the `hello_world()` method is run. What we actually find, is that IPython enters the multi-line mode, so we see the following:
```py
In [2]: hello_world()
..:
```
This is only the case on Windows. On macOS and Unix it works fine.
I'm not entirely sure whether this problem is due to this normalization or due to some IPython behavior. [This issue](https://github.com/ipython/ipython/issues/13054) is already open for it.
Furthermore, because of this issue, the VSCode Python extension had to add a manual timeout, which might fail on slower machines. See their `sendText()` method [here](https://github.com/saponas/vscode-python/blob/1597f84ec29f320fa707355d340282994417f407/src/client/common/terminal/service.ts#L48-L56) and my [summary here](https://github.com/ipython/ipython/issues/13054#issuecomment-2598897658).
```ts
this.terminal!.sendText(`${text}`);
// send carriage return to make code execute
await new Promise(resolve => setTimeout(resolve, 100));
this.terminal.sendText('\u000D', false);
```
| under-discussion | low | Major |
2,796,502,077 | flutter | Engine git-hooks are executed even for framework-only PRs | ```sh
matanl@matanl-macbookpro4 flutter % git push -f
The clang-tidy check is disabled. To enable set the environment variable PRE_PUSH_CLANG_TIDY to any value.
Starting formatting checks.
```
Not only is this taking forever (I believe unrelated), there is nothing it could be checking in this PR. | team-infra,P2,monorepo | low | Minor |
2,796,512,009 | ant-design | Buttons with variant "link" or "text" use the `info` color when the color is set to `primary` | ### Reproduction link
[](https://stackblitz.com/edit/react-t3ejzn?file=demo.tsx)
### Steps to reproduce
- Customize primary color and info color
- Use a button and pass the props color="primary" variant="text"
### What is expected?
The text color should be the primary color
### What is actually happening?
The text color is the info color
| Environment | Info |
| --- | --- |
| antd | 5.23.1 |
| React | doesn't matter |
| System | doesn't matter |
| Browser | doesn't matter |
---
So it is basically the same issue as https://github.com/ant-design/ant-design/issues/51343.
I know you closed it but I really want you to reconsider it.
With the new colors of buttons (which released in version `5.23.0`) the `text` variant is working great.
I think that if you explicitly set the button color to `primary` it must have a primary color.
<!-- generated by ant-design-issue-helper. DO NOT REMOVE --> | unconfirmed | low | Minor |
2,796,519,893 | flutter | Use `DEPS` to bootstrap the GitHub Actions Dart SDK | Some pondering on if initializing the Dart SDK can be vastly simplified, potentially also removing the need for the engine to create a copy of the Dart SDK as part of vended artifacts. One of the reasons this _could_ be beneficial is we'd like to explore using GitHub actions in the Flutter repo.
A GitHub action would _not_ wait for the engine to be built, therefore it would not have access to the engine binaries for the current SHA, and for the action to be written in Dart (instead of say, Python or Bash), it would need a version of the Dart SDK from somewhere - and ideally that would be the same version as pinned in Flutter for compatibility reasons.
In https://github.com/flutter/flutter/pull/161788 I experiment with an override, `FLUTTER_ENGINE_VERSION`, that allows the GitHub action to use the Dart SDK from the base-ref (which merged commit the PR was branched off originally). This will work _until the Dart SDK revs_, in which case now we'll be using an older Dart SDK to evaluate the Flutter repo.
Potential issues: what if we wanted to run `dart format` on a GitHub action? We'd be using an older version of Dart format than the one potentially being changed/updated in `DEPS`.
The proposal below is an _idea_ of what we could do, but we could imagine other ideas/overrides, for example, we could just have the GitHub _action_ self-bootstrap the Dart SDK from `DEPS` and _not_ change how `bin/dart` works at all.
---
Today, `bin/dart` (part of the tools shipped in the Flutter SDK) is a wrapper script that self-bootstraps:
https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/dart#L46-L59
It, if necessary, downloads the Dart SDK, by writing a `engine.version` file:
https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/internal/update_engine_version.sh#L41
... and then using that value in a later shell script to download a Dart SDK:
https://github.com/flutter/flutter/blob/d59d02c014428ac71eb347165d4616043354b7f4/bin/internal/shared.sh#L147
https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/internal/update_dart_sdk.sh#L126-L161
Lots of this logic is due to the fact that the engine and framework repos were separate, there was no (canonical) way to identify which Dart SDK to use, but now there is a single canonical number that is used:
https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/DEPS#L59 | team-infra,P3,c: tech-debt,monorepo | low | Major |
2,796,520,756 | angular | ⚠️ Local development bugs with HMR | ### Which @angular/* package(s) are the source of the bug?
compiler-cli
### Is this a regression?
Yes
### Description
We're aware of reports of local development bugs arising from Angular template HMR in 19.1.x, which has been enabled by default.
List of known issues:
* [Fixed in 19.1.2] ~~#59581~~
* [Fixed in 19.1.2] ~~#59592~~
* [Fixed in 19.1.2] ~~angular/angular-cli#29248~~
* [Fixed in 19.1.3(CLI)] ~~#59605~~
* [Fixed in 19.1.4(CLI)] ~~angular/angular-cli#29424~~
* [Fixed in 19.1.3] ~~#59632~~
* [Fixed in 19.1.3] ~~#59634~~
* [Fixed in 19.1.3] ~~#59639~~
* [Fixed in 19.1.3] ~~#59640~~
We're actively investigating and fixing these issues to create a smooth HMR experience out of the box. Production builds and releases are not affected. If you discover any additional HMR challenges which don't appear to be related to the linked issues above, verify that you are on a recent version (19.1.3+), and please [file a new issue](http://github.com/angular/angular/issues/) and link it here so we can triage and investigate.
### Workarounds
In the meantime, if you encounter any problems relating to HMR, you can work around them in three ways:
1. Refresh the page \- This loads a fresh copy of your application with up to date code.
2. `ng serve --no-hmr` \- This disables HMR for your local devserver and automatically refreshes the page on any code changes.
3. Disable HMR for your project \- Configuring `angular.json` will disable HMR for all `ng serve` invocations by default.
* This is useful for a project with a large number of developers without needing to communicate these workarounds to all contributors.
* Remember to re-enable once the relevant issues are fixed\!
```json
{
"projects": {
"my-project": {
"architect": {
"serve": {
"options": {
"hmr": false
}
}
}
}
}
}
```
Thank you for sharing any HMR issues and helping improve Angular one bugfix at a time. We're committed to fully stabilizing this feature and excited to bring improved build performance to everyone. | area: core,core: hot module replacement (HMR) | medium | Critical |
2,796,528,295 | flutter | A label to track approval of `c: proposal` issues? | For example, in https://github.com/flutter/flutter/issues/161843, I propose making a change.
It would be nice for us to mark proposals as approved, even if it's just a GitHub label, so that folks can know "this is a proposal, and the domain owner(s) agree with it being implemented" versus "this is a proposal, and we haven't reviewed/rejected/approved it yet". | team,will need additional triage,c: proposal,fyi-infra | low | Minor |
2,796,545,097 | rust | Delay stabilizing async closures to consider if they should return `impl IntoFuture` instead of `impl Future` | Async closures are currently in beta for stabilization in 1.85. I'm very excited to see this feature stabilized!
That said, I would like the Lang team to consider delaying their stabilization to consider a new design detail. Delaying stabilization would amount to putting a PR in nightly that undoes the stabilization, and then doing a beta backport.
Note that this doesn't necessarily mean the lang team would be committing to making this change, but if the lang team feels there's a non-trivial chance they would prefer the `IntoFuture` semantics, then we should delay stabilization so there is time to consider this.
I talked with @tmandry and @traviscross about this and they did not remember discussing this particular detail during the stabilization process. If we are mistaken and this was discussed, then we should let the decision stand and not change anything.
# Why should T-lang consider this change?
@yoshuawuyts recently published a [blog post about generators and auto trait leakage](https://blog.yoshuawuyts.com/gen-auto-trait-problem/), which argues that `gen {}` should return `impl IntoIterator` instead of `impl Iterator`. The key example in the post is (adapted slightly):
```rust
let iter = gen {
let rc = Rc::new(...);
yield 12u32;
rc.do_something();
};
spawn(|| {
for num in iter {
println!("{num}");
}
}).unwrap();
```
If `gen {}` returns `impl Iterator`, as it does on nightly today, this example fails to compile because the `Rc` is not `Send`. However, this is overly restrictive because the `Rc` is not created until the spawned thread calls `next()` the first time.
Instead, having `gen {}` return `impl IntoIterator` allows this to work, because the `IntoIterator` part could be `Send`, and once it becomes an `Iterator` it would no longer be `Send`.
This has a similar shape to annoyances we've seen with async Rust, where it's not uncommon for people to want to write futures that are `Send` until they are polled for the first time. If we had made `async {}` blocks return `impl IntoFuture` instead, this would be straightforward to do for the same reasons.
So then the next logical question is what should generator closures or async closures do? For much the same reason, it would be nice for async closures to return futures that are `Send` until first polled.
At this point, changing the return type for `async {}` would require an edition change. However, we have a brief window where we can still make this change for async closures.
# Why should T-lang keep the status quo?
A lot of care has gone into the current design by @compiler-errors and others to come up with something that balances a lot of competing goals. For example, with the current design, the following example works, but it would break with this proposed change:
```rust
fn foo<F, U>(_f: F)
where
F: Fn() -> U,
U: Future
{}
// can be invoked with legacy, fake async closures or new, real async closures
foo(|| async {});
foo(async || {});
```
This is a desirable property, since a lot of existing ecosystem code is written using `Fn() -> impl Future` as an approximation of an async closure. If async closures returned `IntoFuture` then they would not work until existing code is updated to support async closures natively.
I bring this up mainly to show that it's not obvious that we actually should make this change. There are many factors in tension and there's a good chance we already have the optimal balance. We'd want to thoroughly consider any changes, so this issue is mostly about whether we want to take the time to consider those changes.
# Mitigating Factors
This would still be changeable over an edition if we were to decide it's worthwhile but don't want to churn things now. If we were to make this change, we'd likely want to change the behavior of `async {}`. Doing both over an edition at the same time would keep the overall language more consistent. | T-lang,C-discussion,I-lang-nominated | medium | Critical |
2,796,550,635 | godot | Editor crash while refactoring scripts | ### Tested versions
- Occurred in v4.4.dev7.official [46c8f8c5c]
### System information
Godot v4.4.dev7 - Windows 11 (build 22631) - Multi-window, 1 monitor - OpenGL 3 (Compatibility) - NVIDIA GeForce RTX 3080 Ti (NVIDIA; 32.0.15.6636) - AMD Ryzen 9 5900X 12-Core Processor (24 threads)
### Issue description
I experienced two crashes while performing some script/class refactoring in the Godot editor. The exact repro steps are unknown, however, I have two crash dumps from Windows. Both are double frees, so my best guess is an issue with saving and compiling scripts sharing resources improperly.
[Godot_v4.4-dev7_win64.exe.10648.dmp](https://github.com/user-attachments/files/18462120/Godot_v4.4-dev7_win64.exe.10648.dmp)
[Godot_v4.4-dev7_win64.exe.25672.dmp](https://github.com/user-attachments/files/18462119/Godot_v4.4-dev7_win64.exe.25672.dmp)
### Steps to reproduce
Some combination of the following actions in quick succession in an unknown order: (e.g. when duplicating `wall_*.gd` for a new type)
- Ctrl+D (duplicate) a script with a `class_name`
- Double click to open the script
- Ctrl+S to save the script
- Updating the `class_name {name}` name
### Minimal reproduction project (MRP)
Sample project to bootstrap reproduction
[poc.zip](https://github.com/user-attachments/files/18462323/poc.zip) | bug,topic:gdscript,topic:editor,crash | low | Critical |
2,796,555,929 | rust | LLVM assertion failure on Wasm exceptions | ### Code
```Rust
fn main() {}
```
:/
### Meta
I'm building rustc on the current master (bcd0683e5dce1945b5d940714742e7502883bb5c) with LLVM assertions enabled. I'm compiling the program above like this:
```
RUSTFLAGS="-C panic=unwind" cargo +stage1 build --target wasm32-wasip1 -Z build-std
```
I'm using the 23.0 WASI sysroot, downloaded from their GitHub releases, if that matters.
### Error output
```
rustc: /checkout/src/llvm-project/llvm/lib/CodeGen/SelectionDAG/SelectionDAGISel.cpp:1378: void mapWasmLandingPadIndex(MachineBasicBlock *, const CatchPadInst *): Assertion `IntrFound && "wasm.landingpad.index intrinsic not found!"' failed.
```
like, everywhere, on all sorts of standard crates like `libc` and `unwind`.
Disabling assertions hides the bug, and user code works fine. This also does not reproduce on rustup's nightly, probably for the same reason. I have no idea, really. Please tell me I'm doing something wrong.
For context, this seems more important because Emscripten is supposed to switch to Wasm exceptions soon-ish, and I imagine the same problem occurs there (but I haven't checked). Unlike wasm32-wasip1, unwinding is enabled there by default.
@rustbot label +A-LLVM | I-crash,A-LLVM,T-compiler,O-wasm,C-bug,O-wasi,O-emscripten | medium | Critical |
2,796,566,382 | ollama | HSA_OVERRIDE_GFX_VERSION_0 while running on only one GPU | ### What is the issue?
I am running `ollama:rocm` in a docker container on Ubuntu 24.04. My GPU is an RX 6600 (`gfx1032`). Everything works fine when I run `ollama` using
```bash
docker run -d \
--device /dev/kfd \
--device /dev/dri \
-v ollama:/root/.ollama \
-p 11434:11434 \
--restart unless-stopped \
--env HSA_OVERRIDE_GFX_VERSION="10.3.0" \
--env OLLAMA_KEEP_ALIVE="-1" \
--name ollama-rocm \
ollama/ollama:rocm
```
```
time=2025-01-18T01:24:54.470Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.5-0-g32bd37a-dirty)"
time=2025-01-18T01:24:54.470Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu_avx2 rocm_avx cpu cpu_avx]"
time=2025-01-18T01:24:54.470Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
time=2025-01-18T01:24:54.472Z level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory"
time=2025-01-18T01:24:54.472Z level=INFO source=amd_linux.go:391 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION=10.3.0
time=2025-01-18T01:24:54.472Z level=INFO source=types.go:131 msg="inference compute" id=0 library=rocm variant="" compute=gfx1032 driver=0.0 name=1002:73ff total="8.0 GiB" available="8.0 GiB"
```
However, when I run using
```bash
docker run -d \
--device /dev/kfd \
--device /dev/dri \
-v ollama:/root/.ollama \
-p 11434:11434 \
--restart unless-stopped \
--env HSA_OVERRIDE_GFX_VERSION_0="10.3.0" \
--env OLLAMA_KEEP_ALIVE="-1" \
--name ollama-rocm \
ollama/ollama:rocm
```
`ollama` runs only on CPU:
```
time=2025-01-18T01:25:58.373Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.5-0-g32bd37a-dirty)"
time=2025-01-18T01:25:58.373Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu_avx2 rocm_avx cpu cpu_avx]"
time=2025-01-18T01:25:58.373Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
time=2025-01-18T01:25:58.375Z level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory"
time=2025-01-18T01:25:58.378Z level=WARN source=amd_linux.go:378 msg="amdgpu is not supported (supported types:[gfx1030 gfx1100 gfx1101 gfx1102 gfx900 gfx906 gfx908 gfx90a gfx940 gfx941 gfx942])" gpu_type=gfx1032 gpu=0 library=/usr/lib/ollama
time=2025-01-18T01:25:58.378Z level=WARN source=amd_linux.go:385 msg="See https://github.com/ollama/ollama/blob/main/docs/gpu.md#overrides for HSA_OVERRIDE_GFX_VERSION usage"
time=2025-01-18T01:25:58.378Z level=INFO source=amd_linux.go:404 msg="no compatible amdgpu devices detected"
time=2025-01-18T01:25:58.378Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered"
time=2025-01-18T01:25:58.378Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="62.7 GiB" available="61.2 GiB"
```
I am trying this because eventually I want to get newer AMD GPUs and use them concurrently for `ollama`.
Is this not supported when running `ollama` in `docker`?
If I want to use this RX 6600 and an RX 7800 in the same system, how should I do it with `docker`?
### OS
Linux
### GPU
AMD
### CPU
Intel
### Ollama version
0.5.5-0-g32bd37a-dirty | bug | low | Critical |
2,796,575,411 | ollama | ollama/ollama:rocm not detecting AMD GPU being passed in | ### What is the issue?
# Overview
I've been trying to add ollama/openwebui to my home lab and I've run across an issue I cannot figure out. ROCM has finally been released last month for my linux distro and I've got it installed and can run (on gpu) ollama on my host machine. When I try to transition to docker however, I am unable to get it to work. I will provide what I have but please ask any clarifying questions.
## Components
Host: Debian 6.1.124-1
CPU: I7-12700K
GPU: Radeon 6900 XT
# Docker compose file
```
services:
ollama:
image: ollama/ollama:rocm
container_name: ollama
hostname: ollama
privileged: true
volumes:
- ollama:/root/.ollama
environment:
HSA_OVERRIDE_GFX_VERSION: 10.3.0
AMD_SERIALIZE_KERNEL: 3
HIP_VISIBLE_DEVICES: 0
OLLAMA_DEBUG: 1
AMD_LOG_LEVEL: 3
# ports:
# - :11434
restart: unless-stopped
networks:
- npm
devices:
- /dev/kfd
- /dev/dri
group_add:
- 105
- 44
security_opt:
- seccomp:unconfined
```
# Logs & Outputs
No errors that I can really tell. Here are the logs on startup in docker
```
2025-01-17 20:40:33 2025/01/18 01:40:33 routes.go:1187: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES:0 HSA_OVERRIDE_GFX_VERSION:10.3.0 HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:true OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.654Z level=INFO source=images.go:432 msg="total blobs: 0"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=images.go:439 msg="total unused blobs removed: 0"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.7-0-ga420a45-dirty)"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:80 msg="runners located" dir=/usr/lib/ollama/runners
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx/ollama_llama_server
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx2/ollama_llama_server
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/rocm_avx/ollama_llama_server
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu cpu_avx cpu_avx2 rocm_avx]"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=routes.go:1268 msg="Override detection logic by setting OLLAMA_LLM_LIBRARY"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=sched.go:105 msg="starting llm scheduler"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:99 msg="searching for GPU discovery libraries for NVIDIA"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcuda.so*
2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/lib/ollama/libcuda.so* /libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths=[]
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcudart.so*
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /libcudart.so* /usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths=[]
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=amd_linux.go:421 msg="amdgpu driver not detected /sys/module/amdgpu"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered"
2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="50.7 GiB" available="35.1 GiB"
```
```ls -ld /sys/module/amdgpu``` is display ```drwxr-xr-x 7 root root 0 Jan 17 12:23 /sys/module/amdgpu```
### OS
Linux, Docker
### GPU
AMD
### CPU
Intel
### Ollama version
0.5.7-0-ga420a45-dirty | bug | low | Critical |
2,796,582,590 | tauri | linux development often show blank screen. | when modify some code , sometimes the `webview` component show nothing.
I need restart `pnpm run dev` and restart the rust `backend` to make it working again.
Anyone else have the same problem??? | type: bug,platform: Linux,status: needs triage | low | Major |
2,796,586,042 | PowerToys | Technical glitches while running the powertoy | ### Microsoft PowerToys version
0.87.1.0
### Installation method
Microsoft Store
### Running as admin
Yes
### Area(s) with issue?
PowerToys Run, General
### Steps to reproduce
when my pc is in idle state and after some time when i try to run the powertoys run using alt + space it don't register the action and fail to open
### ✔️ Expected Behavior
when i open my pc with in 10 sec maximum it should open the powertoys run
### ❌ Actual Behavior
It does not open the powertoys run and it some time it open directly using space button , also experienced at the same time that the it is unable to register my keyboard input and no keyword input seen
sorry i don't have the screenshots right now
### Other Software
_No response_ | Issue-Bug,Needs-Triage | low | Minor |
2,796,598,858 | rust | Tracking issue for release notes of #133700: const-eval: detect more pointers as definitely not-null |
This issue tracks the release notes text for #133700.
### Steps
- [ ] Proposed text is drafted by PR author (or team) making the noteworthy change.
- [ ] Issue is nominated for release team review of clarity for wider audience.
- [ ] Release team includes text in release notes/blog posts.
### Release notes text
The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing).
````markdown
# Language
- [const-eval: detect more pointers as definitely not-null based on their alignment](https://github.com/rust-lang/rust/pull/133700)
````
> [!TIP]
> Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use.
> The category will be de-duplicated with all the other ones by the release team.
>
> *More than one section can be included if needed.*
### Release blog section
If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section.
*Otherwise leave it empty.*
````markdown
````
cc @RalfJung, @lcnr -- origin issue/PR authors and assignees for starting to draft text
| T-lang,T-compiler,relnotes,needs-triage,relnotes-tracking-issue | low | Minor |
2,796,602,141 | neovim | AI/LLM/GenAI related features | This is a stub issue, will update it over time.
## Problem
Although "AI features" are not something we see a need for in Nvim core, I'm looking at the space and have noticed some "primitives" that are missing and require every plugin to reimplement things that are generally useful, even for non-AI plugins.
## Expected behavior
- A way to mark a buffer or window as "busy" or "in progress".
- Default 'statusline' shows the busy indicator (optionally progress "spinner"?).
- https://github.com/neovim/neovim/issues/28809
- `vim.ui` interface for showing progress.
- ✅ blink/highlight of a region will be helpful https://github.com/neovim/neovim/pull/32012 cc @siddhantdev
- multiline completion https://github.com/neovim/neovim/issues/7769 | ui,lua | low | Major |
2,796,603,367 | godot | [godot3.6] Compile web platform runtime error | ### Tested versions
3.6
### System information
godot3.6 v3.6.1.rc.mono.custom_build [4c1514c7f]
### Issue description
compile for the Web platform
`scons platform=javascript tools=no target=release
scons platform=javascript tools=no target=release_debug
`
set compiled webassembly_debug.zip and webassembly_release.zip

run web got a error
modules\mono\glue\gd_glue.cpp:249 - Failed to export project: Missing BCL (Base Class Library) for platform: javascript


### Steps to reproduce
-
### Minimal reproduction project (MRP)
- | bug,platform:web,needs testing,topic:dotnet | low | Critical |
2,796,610,756 | PowerToys | Not able to copy certain info | ### Microsoft PowerToys version
0.87.1
### Installation method
PowerToys auto-update
### Running as admin
Yes
### Area(s) with issue?
TextExtractor
### Steps to reproduce
Tried to copy info. Text Extractor failed to copy.

### ✔️ Expected Behavior
Successful copy
### ❌ Actual Behavior
Failed to copy particular info. Copied everything else on the page successfully.
### Other Software
COMODO - Internet Security - 12.2.2.7036 | Issue-Bug,Needs-Triage,Product-Text Extractor | low | Critical |
2,796,611,015 | react | [DevTools Bug] Cannot add child "301" to parent "155" because parent node was not found in the Store. | ### Website or app
www.github.com
### Repro steps
The error was thrown at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1173126
at v.emit (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1140783)
at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1142390
at bridgeListener (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1552662)
### How often does this bug happen?
Every time
### DevTools package (automated)
react-devtools-extensions
### DevTools version (automated)
6.0.1-c7c68ef842
### Error message (automated)
Cannot add child "301" to parent "155" because parent node was not found in the Store.
### Error call stack (automated)
```text
at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1173126
at v.emit (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1140783)
at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1142390
at bridgeListener (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1552662)
```
### Error component stack (automated)
```text
```
### GitHub query string (automated)
```text
https://api.github.com/search/issues?q=Cannot add child to parent because parent node was not found in the Store. in:title is:issue is:open is:public label:"Component: Developer Tools" repo:facebook/react
``` | Type: Bug,Status: Unconfirmed,Component: Developer Tools | medium | Critical |
2,796,615,965 | godot | [v4.4-dev7_win64] Failure to launch | ### Tested versions
Reproducible in: Godot_v4.4-dev7_win64, Godot_v4.4-beta1_win64
Not reproducible in: Godot_v4.4-dev6_win64
### System information
Windows 11 - RTX 4090 (Thunderbolt 4) - Driver 566.36
### Issue description
Upon running the Godot application, the cursor spins for a second, then stops without ever launching the application. When utilizing onboard graphics (Radeon 680M), this problem is not present. However, when utilizing an RTX 4090 over Thunderbolt 4 (eGPU), it occurs. Just to clarify, this is a full-size card, not a 4090 mobile.
Logging doesn't appear out of the ordinary to me. If there is additional logging or information I can gather to assist, please do share.
```
PS C:\Users\kaboo\Desktop> .\Godot_v4.4-dev7_win64_console.exe --verbose
WorkerThreadPool: 16 threads, 4 max low-priority.
Godot Engine v4.4.dev7.official.46c8f8c5c - https://godotengine.org
TextServer: Added interface "Dummy"
TextServer: Added interface "ICU / HarfBuzz / Graphite (Built-in)"
Native OpenGL API detected: 3.3: NVIDIA - NVIDIA GeForce RTX 4090
NVAPI: Init OK!
NVAPI: Disabled OpenGL threaded optimization successfully
NVAPI: Disabled G-SYNC for windowed mode successfully
PS C:\Users\kaboo\Desktop>
```
### Steps to reproduce
Appears to possibly be hardware related. Steps are as simple as launching the application.
### Minimal reproduction project (MRP)
Not Applicable | bug,platform:windows,crash,regression | low | Critical |
2,796,626,997 | rust | regression: ICE: Failed to normalize Alias(Opaque, AliasTy | - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/Menestis.Skynet/log.txt
```
[INFO] [stdout] error: internal compiler error: compiler/rustc_middle/src/ty/normalize_erasing_regions.rs:168:90: Failed to normalize Alias(Opaque, AliasTy { args: [], def_id: DefId(0:850 ~ skynet[716d]::web::create_task::{opaque#0}::{opaque#0}), .. }), maybe try to call `try_normalize_erasing_regions` instead
```
| I-ICE,P-low,T-compiler,regression-from-stable-to-beta | low | Critical |
2,796,627,464 | rust | regression: a value of type `HashMap<Pulse, u64>` cannot be built from | - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/dimo414.advent-2023/log.txt
```
[INFO] [stderr] Compiling advent-2023 v0.1.0 (/opt/rustwide/workdir)
[INFO] [stdout] error[E0277]: a value of type `HashMap<Pulse, u64>` cannot be built from an iterator over elements of type `&(Pulse, {integer})`
[INFO] [stdout] --> src/bin/20/main.rs:278:88
[INFO] [stdout] |
[INFO] [stdout] 278 | assert_eq!(config.counts, [(Pulse::Low, low), (Pulse::High, high)].into_iter().collect());
[INFO] [stdout] | ^^^^^^^ value of type `HashMap<Pulse, u64>` cannot be built from `std::iter::Iterator<Item=&(Pulse, {integer})>`
[INFO] [stdout] 279 | } }
``` | T-lang,T-compiler,regression-from-stable-to-beta,I-prioritize,I-lang-nominated,A-edition-2024 | low | Critical |
2,796,627,814 | rust | regression: cannot borrow ... as immutable because it is also borrowed as mutable | - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/LP-Finance-Inc.twamm/log.txt
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/Mokosha.pbrt_rust/log.txt
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/jw1912.diffable/log.txt
```
[INFO] [stdout] error[E0502]: cannot borrow `*inputs` as immutable because it is also borrowed as mutable
[INFO] [stdout] --> examples/basic.rs:267:27
[INFO] [stdout] |
[INFO] [stdout] 263 | if let Some(grd) = inputs[0].grad.as_mut() {
[INFO] [stdout] | -------------- mutable borrow occurs here
[INFO] [stdout] 264 | / *grd += output_grad
[INFO] [stdout] 265 | | * if inputs[0].val > 0.0 {
[INFO] [stdout] 266 | | 1.0
[INFO] [stdout] 267 | | } else if inputs[0].val == 0.0 {
[INFO] [stdout] | | ^^^^^^^^^ immutable borrow occurs here
[INFO] [stdout] ... |
[INFO] [stdout] 270 | | -1.0
[INFO] [stdout] 271 | | };
[INFO] [stdout] | |_________________- mutable borrow later used here
``` | T-compiler,regression-from-stable-to-beta,P-critical,T-types | low | Critical |
2,796,627,979 | rust | regression: ICE: `trimmed_def_paths` called, diagnostics were expected but none were emitted | - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/reg/quinn-proto-0.11.9/log.txt
```
[INFO] [stderr] `trimmed_def_paths` called, diagnostics were expected but none were emitted. Use `with_no_trimmed_paths` for debugging. This happened in the following `must_produce_diag` call's backtrace:
[INFO] [stderr] 0: <rustc_errors::DiagCtxtHandle>::set_must_produce_diag
[INFO] [stderr] 1: <rustc_session::session::Session>::record_trimmed_def_paths
[INFO] [stderr] 2: rustc_middle::ty::print::pretty::trimmed_def_paths
[INFO] [stderr] 3: rustc_query_impl::plumbing::__rust_begin_short_backtrace::<rustc_query_impl::query_impl::trimmed_def_paths::dynamic_query::{closure#2}::{closure#0}, rustc_middle::query::erase::Erased<[u8; 8]>>
[INFO] [stderr] 4: rustc_query_system::query::plumbing::try_execute_query::<rustc_query_impl::DynamicConfig<rustc_query_system::query::caches::SingleCache<rustc_middle::query::erase::Erased<[u8; 8]>>, false, false, false>, rustc_query_impl::plumbing::QueryCtxt, false>
[INFO] [stderr] 5: rustc_query_impl::query_impl::trimmed_def_paths::get_query_non_incr::__rust_end_short_backtrace
``` | I-ICE,P-high,T-compiler,regression-from-stable-to-beta | low | Critical |
2,796,677,342 | flutter | I get this error in IOS : [ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PlatformException(channel-error, Unable to establish connection on channel: "dev.flutter.pigeon.path_provider_foundation.PathProviderApi.getDirectoryPath"., null, null) #0 | When I try to run this in IOS I get this error.
[ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PlatformException(channel-error, Unable to establish connection on channel: "dev.flutter.pigeon.path_provider_foundation.PathProviderApi.getDirectoryPath"., null, null)
#0 PathProviderApi.getDirectoryPath (package:path_provider_foundation/messages.g.dart:95:7) | waiting for customer response,in triage | low | Critical |
2,796,689,807 | deno | Relax restrictions on scrypt `log_n` | Version: Deno 2.1.6
```js
// scrypt.js
import crypto from 'node:crypto'
crypto.scrypt('password', 'salt', 128, (err, hashed) => {
if (err) throw err
console.log('hashed = ', hashed.toString('hex'))
})
crypto.scrypt('password', 'salt', 128, (err, hashed) => {
if (err) throw err
console.log('hashed = ', hashed.toString('hex'))
})
```
```shell
$ deno ./scrypt.js
============================================================
Deno has panicked. This is a bug in Deno. Please report this
at https://github.com/denoland/deno/issues/new.
If you can reliably reproduce this panic, include the
reproduction steps and re-run with the RUST_BACKTRACE=1 env
var set and include the backtrace in your report.
Platform: linux x86_64
Version: 2.1.6
Args: ["/home/cyan/.asdf/installs/deno/2.1.4/bin/deno", "./scrypt.js"]
thread 'tokio-runtime-worker' panicked at ext/node/ops/crypto/mod.rs:585:4:
called `Result::unwrap()` on an `Err` value: InvalidParams
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
============================================================
Deno has panicked. This is a bug in Deno. Please report this
at https://github.com/denoland/deno/issues/new.
If you can reliably reproduce this panic, include the
reproduction steps and re-run with the RUST_BACKTRACE=1 env
var set and include the backtrace in your report.
Platform: linux x86_64
Version: 2.1.6
Args: ["/home/cyan/.asdf/installs/deno/2.1.4/bin/deno", "./scrypt.js"]
thread 'tokio-runtime-worker' panicked at ext/node/ops/crypto/mod.rs:585:4:
called `Result::unwrap()` on an `Err` value: InvalidParams
```
```shell
$ node ./scrypt.js
hashed = 745731af4484f323968969eda289aeee005b5903ac561e64a5aca121797bf7734ef9fd58422e2e22183bcacba9ec87ba0c83b7a2e788f03ce0da06463433cda64176095fbbad7dc98c33fd75955b4b29c94f6e97617bd68d8ff17cf1ed5ad12f3fc6c8eb5b844f2d003ebaf5eaed19e8f665928472a6941f7efc6ebcdd6fd13a
hashed = 745731af4484f323968969eda289aeee005b5903ac561e64a5aca121797bf7734ef9fd58422e2e22183bcacba9ec87ba0c83b7a2e788f03ce0da06463433cda64176095fbbad7dc98c33fd75955b4b29c94f6e97617bd68d8ff17cf1ed5ad12f3fc6c8eb5b844f2d003ebaf5eaed19e8f665928472a6941f7efc6ebcdd6fd13a
```
Here:
https://github.com/denoland/deno/blob/57dd66ec3dae2a0fe376f6a43c476dfade421c84/ext/node/ops/crypto/mod.rs#L579-L585
The Result is unwrapped, and it will be an `Err` when len > 64
According to the Document here: https://docs.rs/scrypt/latest/scrypt/struct.Params.html#conditions
But the same thing works in Node.js, maybe deno need to consider some alternatives? | upstream,node compat,crypto | low | Critical |
2,796,720,352 | godot | Duplicating resources and editing them without saving stores incorrect data in them and inspector displays wrong data. | ### Tested versions
Reproducible in:
- v4.4.beta1.official [d33da79d3]
- v4.3.stable.official [77dcf97d8]
- v4.2.2.stable.official [15073afe3]
### System information
Godot v4.3.stable - Windows 10.0.19045 - GLES3 (Compatibility) - NVIDIA GeForce GTX 750 Ti (NVIDIA; 32.0.15.6094) - Intel(R) Core(TM) i5-6400 CPU @ 2.70GHz (4 Threads)
### Issue description
When setting up a resource, then duplicating that resource an setting it up (a few times) the resource file will have icorrect data in it but display the correct data in the editor. When the project is executed, the incorrect information will be used.
https://github.com/user-attachments/assets/4ab695f9-bf9a-4c32-a93a-72d2f485346a
### Steps to reproduce
1. Create a Shortcut resource.
2. Assign it an InputKeyEvent
3. Duplicate that resource
4. Edit that resource InputKeyEvent to be something new.
5. (..repeat 2 or 3 times)
6. Run project.
Shortcuts are assigned wrongly
**Note:** When creating the shortcuts directly (Create New -> Resource) this didn't happen. Issue only occours when duplicating (Ctrl+D).
The Minimal Reproduction Project can't be uploaded in its entirety since the editor will display the correctly stored data on reload.
### Minimal reproduction project (MRP)
[shortcut_saving.zip](https://github.com/user-attachments/files/18462889/shortcut_saving.zip) | topic:editor,needs testing | low | Minor |
2,796,746,011 | tensorflow | TFnode on TensorflowonSpark 2.2.5 | I'm running some code on Microsoft Fabric, and I use the following line:
def map_fun(tf_args, ctx):
cluster, server = TFNode.start_cluster_server(ctx)
print('ctx')
if ctx.job_name == "ps":
server.join()
else:
print("Hello from worker", ctx.task_index)
I'm getting an error regarding TFnode, does anyone know why this could be?
TensorFlow version:2.12
TensorflowonSpark:2.2.5
The idea is to print a hello world to see if they are running on differents clusters. | type:support,TF 2.12 | low | Critical |
2,796,791,671 | ollama | Embedding Model: iamgroot42/rover_nexus | Its the top model in MTEB Leaderboard. | model request | low | Minor |
2,796,826,624 | terminal | First turn off the performance option Show window contents when dragging. Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position. | ### Windows Terminal version
1.21.3231.0
### Windows build number
10.0.26100.2894
### Other Software
_No response_
### Steps to reproduce
First turn off the performance option Show window contents when dragging. Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position.
### Expected Behavior
No black line appears when the window is zoomed out.
### Actual Behavior
Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position.
| Issue-Bug,Needs-Author-Feedback,Needs-Triage,No-Recent-Activity | low | Major |
2,796,831,254 | godot | play_backwards Cannot Play Loop Animations | ### Tested versions
v4.4.beta1.official [4ce466d7f]
### System information
Windows 11 (build 22631) - Multi-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2050 (NVIDIA; 32.0.15.6109) - 12th Gen Intel(R) Core(TM) i5-12500H (16 threads)
### Issue description
I have encountered an issue where the `play_backwards()` method cannot play loop animations for models imported from Blender. When attempting to play these animations in reverse, they get stuck on the first frame, while all non-looping animations play correctly in reverse.
### Steps to reproduce
Play a looped animation backwards
### Minimal reproduction project (MRP)
N/A | bug,needs testing,topic:animation | low | Minor |
2,796,841,741 | terminal | Unable to Launch Windows Terminal via `wt.exe` or `windowsterminal.exe` After KB5050021 Update | ### Windows Terminal version
1.21.3231.0
### Windows build number
22621.4751
### Other Software
_No response_
### Steps to reproduce
## Detailed Description
After applying the KB5050021 update on January 14, 2025, the Windows Terminal fails to launch when executed via:
- `wt.exe` in PowerShell, CMD, or VS Code integrated terminal.
- Direct execution of `windowsterminal.exe` in `C:\Program Files\WindowsApps\Microsoft.WindowsTerminal_1.21.3231.0_x64__8wekyb3d8bbwe`.
No error messages are displayed, and the process is not visible in `Task Manager` for long. The `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds. However, the terminal opens correctly when launched via its app icon in the Start menu.
---
## Steps to Reproduce
1. Update Windows 11 to build 22621.4751 via KB5050021.
2. Attempt to open Windows Terminal via `wt.exe` in a PowerShell or CMD window.
3. Alternatively, attempt to open `windowsterminal.exe` directly.
4. Observe that no terminal appears and no error messages are shown.
---
## Observations
- When launched via the app icon, the terminal works correctly.
- Running `Get-Process -Name *WindowsTerminal*, wt` after opening the terminal via the app icon shows `WindowsTerminal` as active. Running the same command after executing `wt.exe` returns no processes.
- When executing `wt.exe`, the `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds.
- The issue persists across the following environments:
- Default PowerShell terminal.
- CMD.
- VS Code integrated terminal.
- Resetting, uninstalling, and reinstalling Windows Terminal, as well as running `sfc /scannow` and DISM commands, did not resolve the issue.
---
## Expected Behavior
`wt.exe` and `windowsterminal.exe` should launch Windows Terminal normally, as before the update.
## Additional Notes
The issue affects workflows that depend on scripting or terminal automation via `wt.exe`. This problem has been consistently reproducible and may affect other users who installed the KB5050021 update.
### Expected Behavior
## Expected Behavior
`wt.exe` and `windowsterminal.exe` should launch Windows Terminal normally, as before the update.
---
### Actual Behavior
No error messages are displayed, and the process is not visible in `Task Manager` for long. The `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds. However, the terminal opens correctly when launched via its app icon in the Start menu. | Issue-Bug,Needs-Triage,Needs-Attention | low | Critical |
2,796,865,708 | storybook | [Bug]: create Nextjs sandbox failed. | ### Describe the bug
```
⚙️ Initializing Storybook
> node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug
╭──────────────────────────────────────────────────────────────╮
│ │
│ Adding Storybook version 8.6.0-alpha.0 to your project.. │
│ This is a pre-release version. │
│ │
╰──────────────────────────────────────────────────────────────╯
• Detecting project type. ✓
Installing dependencies...
➤ YN0000: · Yarn 4.6.0
➤ YN0000: ┌ Resolution step
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0000: └ Completed
➤ YN0000: ┌ Link step
➤ YN0000: └ Completed
➤ YN0000: · Done in 0s 269ms
• Adding Storybook support to your "Next" app
✔ Getting the correct version of 9 packages
Configuring eslint-plugin-storybook in your package.json
⠧ Installing Storybook dependencies
An error occurred while installing dependencies:
YARN2 error
YN0060: INCOMPATIBLE_PEER_DEPENDENCY
-> @storybook/test is listed by your project with version 8.6.0-alpha.0 (p8d038), which doesn't satisfy what @storybook/react requests (but they have non-overlapping ranges!).
YN0060: INCOMPATIBLE_PEER_DEPENDENCY
-> storybook is listed by your project with version 8.6.0-alpha.0 (p23005), which doesn't satisfy what @storybook/addon-essentials and other dependencies request (but they have non-overlapping ranges!).
YN0002: MISSING_PEER_DEPENDENCY
-> before-storybook@workspace:. doesn't provide webpack (pd8fe4), requested by @storybook/nextjs.
YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA
-> Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA
-> Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
YN0001: EXCEPTION
-> Error: Assertion failed: Writing attempt prevented to /Users/zhyd1997/workspace/storybook/code/addons/essentials/node_modules/@storybook/addon-docs which is outside project root: /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts
Please check the logfile generated at ./storybook.log for troubleshooting and try again.
Error: Error: YARN2 error
YN0060: INCOMPATIBLE_PEER_DEPENDENCY
-> @storybook/test is listed by your project with version 8.6.0-alpha.0 (p8d038), which doesn't satisfy what @storybook/react requests (but they have non-overlapping ranges!).
YN0060: INCOMPATIBLE_PEER_DEPENDENCY
-> storybook is listed by your project with version 8.6.0-alpha.0 (p23005), which doesn't satisfy what @storybook/addon-essentials and other dependencies request (but they have non-overlapping ranges!).
YN0002: MISSING_PEER_DEPENDENCY
-> before-storybook@workspace:. doesn't provide webpack (pd8fe4), requested by @storybook/nextjs.
YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA
-> Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA
-> Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
YN0001: EXCEPTION
-> Error: Assertion failed: Writing attempt prevented to /Users/zhyd1997/workspace/storybook/code/addons/essentials/node_modules/@storybook/addon-docs which is outside project root: /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts
Please check the logfile generated at ./storybook.log for troubleshooting and try again.
at Yarn2Proxy.addDependencies (/Users/zhyd1997/workspace/storybook/code/core/dist/common/index.cjs:25592:15)
at async baseGenerator (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2166:5)
at async generator4 (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2350:3)
at async installStorybook (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2966:12)
at async doInitiate (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:3081:25)
at async withTelemetry (/Users/zhyd1997/workspace/storybook/code/core/dist/core-server/index.cjs:51862:12)
at async initiate (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:3146:26)
An error occurred while executing: `node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug`
🚨 Initializing Storybook failed
Error running task sandbox for nextjs/default-ts:
{
"shortMessage": "Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug",
"command": "node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug",
"escapedCommand": "\"node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug\"",
"exitCode": 1,
"failed": true,
"timedOut": false,
"isCanceled": false,
"killed": false
}
Error: Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug
at makeError (/Users/zhyd1997/workspace/storybook/scripts/node_modules/execa/lib/error.js:59:13)
at handlePromise (/Users/zhyd1997/workspace/storybook/scripts/node_modules/execa/index.js:119:50)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
at async exec (/Users/zhyd1997/workspace/storybook/scripts/utils/exec.ts:43:7)
at async executeCLIStep (/Users/zhyd1997/workspace/storybook/scripts/utils/cli-step.ts:105:3)
at async init (/Users/zhyd1997/workspace/storybook/scripts/tasks/sandbox-parts.ts:145:3)
at async Object.run (/Users/zhyd1997/workspace/storybook/scripts/tasks/sandbox.ts:111:5)
at async runTask (/Users/zhyd1997/workspace/storybook/scripts/task.ts:319:24)
at async run (/Users/zhyd1997/workspace/storybook/scripts/task.ts:489:28) {
shortMessage: 'Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug',
command: 'node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug',
escapedCommand: '"node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug"',
exitCode: 1,
signal: undefined,
signalDescription: undefined,
stdout: undefined,
stderr: undefined,
failed: true,
timedOut: false,
isCanceled: false,
killed: false
}
➜ storybook git:(next) ✗
```
### Reproduction link
NA
### Reproduction steps
1. yarn task
2. select to generate `nextjs-default-ts` sandbox from template.
3. see the errors above.
### System
```bash
Storybook Environment Info:
System:
OS: macOS 15.1.1
CPU: (14) arm64 Apple M3 Max
Shell: 5.9 - /bin/zsh
Binaries:
Node: 18.16.0 - ~/.nvm/versions/node/v18.16.0/bin/node
Yarn: 4.3.0 - ~/.nvm/versions/node/v18.16.0/bin/yarn <----- active
npm: 9.5.1 - ~/.nvm/versions/node/v18.16.0/bin/npm
Browsers:
Chrome: 131.0.6778.266
Safari: 18.1.1
```
### Additional context
```
➜ storybook git:(next) ✗ yarn task
Installing Script Dependencies...
✔ Which task would you like to run? › Create the sandbox from a template (sandbox)
✔ What template would you like to make a sandbox for? › nextjs/default-ts
To run this directly next time, use:
yarn task --task sandbox --template nextjs/default-ts
👉 Selected sandbox: nextjs/default-ts
Task readiness up to sandbox
🟢 install > 🟢 compile > 🟡 sandbox
✔ We need to run all tasks from sandbox onwards, would you like to start from an earlier task? › Create the sandbox from a template (sandbox)
🟢 install > 🟢 compile > 🔄 sandbox
👷 Bootstrapping Template
> node /Users/zhyd1997/workspace/storybook/code/lib/cli-storybook/bin/index.cjs repro nextjs/default-ts --output /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts --no-init --debug
╭──────────────────────────────────────────────────╮
│ │
│ Creating a Storybook 8.6.0-alpha.0 sandbox.. │
│ This is a pre-release version. │
│ │
╰──────────────────────────────────────────────────╯
⚠️ /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts already exists! Overwriting...
🏃 Adding Next.js Latest (Webpack |
```
[storybook.log](https://github.com/user-attachments/files/18463756/storybook.log) | bug,good first issue,help wanted,cli,nextjs | low | Critical |
2,796,865,913 | puppeteer | [Feature]: Keyboard and Pointer Lock API Permission type added | ### Feature description
Can we please get the new chrome v131 Keyboard and Pointer Lock permissions added?
https://pptr.dev/api/puppeteer.permission
Current workaround requires using v23.7.1 for chrome v130 browser.
https://pptr.dev/supported-browsers
https://developer.chrome.com/blog/keyboard-lock-pointer-lock-permission
We are currently unable to test our application on any version of puppeteer above v23.7.1
KeyboardLock and PointerLock were added to devtools-protocol in this commit (r1388031)
https://github.com/ChromeDevTools/devtools-protocol/commit/e3771a854976cb9579b78c03c58bef4d5c63ca93
| feature,P2,chrome,m133 | low | Minor |
2,796,877,194 | tensorflow | Segmentation fault (core dumped) in `RaggedTensorToTensor` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.17
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tf.raw_ops.RaggedTensorToTensor` triggers crash.
### Standalone code to reproduce the issue
```shell
import tensorflow as tf
shape = tf.constant(-1, shape=[], dtype=tf.int64)
values = tf.constant(0, shape=[0], dtype=tf.int32)
default_value = tf.constant(0, shape=[], dtype=tf.int32)
row_partition_tensors = tf.constant([0, 1, 6], shape=[3], dtype=tf.int64)
row_partition_types = ["ROW_SPLITS"]
tf.raw_ops.RaggedTensorToTensor(
shape=shape,
values=values,
default_value=default_value,
row_partition_tensors=[row_partition_tensors],
row_partition_types=row_partition_types)
```
### Relevant log output
```shell
Segmentation fault (core dumped)
``` | type:bug,comp:ops,2.17 | medium | Critical |
2,796,877,858 | ui | [feat]: I would like to request a change to the cookie name used in the sidebar component. | ### Feature description
The current cookie name for the sidebar is as follows:
```
const SIDEBAR_COOKIE_NAME = "sidebar:state";
```
However, this cookie name causes a validation error in msw, making it unusable.
```
TypeError: argument name is invalid
```
In msw, cookie names are validated against the following regular expression:
msw.js
```
var cookieNameRegExp = /^[!#$%&'*+\-.^_`|~0-9A-Za-z]+$/;
if (!cookieNameRegExp.test(name)) {
throw new TypeError("argument name is invalid");
}
```
The issue was resolved when I changed the cookie name to "sidebar-state".
Could you please adjust the cookie name to comply with this validation?
Thank you!
### Affected component/components
Sidebar
### Additional Context
Additional details here...
### Before submitting
- [x] I've made research efforts and searched the documentation
- [x] I've searched for existing issues and PRs | area: request | low | Critical |
2,796,881,366 | vscode | GPU usage fluctuate in VS Code, but stably high in other IDE. |
Type: <b>Performance Issue</b>
I am building a BERT-like model with pytorch, but I encounter a very unstable GPU usage while training my model on VS Code.
I tested the same code and python env in Spyder IDE, everything is fine and GPU usage is stably above 90%. So, it seems that there is something bog down my device and cause the bottleneck.
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Linux x64 6.8.0-51-generic
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Xeon(R) CPU E7-8890 v3 @ 2.50GHz (144 x 0)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off<br>webnn: disabled_off|
|Load (avg)|0, 1, 12|
|Memory (System)|125.72GB (118.82GB free)|
|Process Argv|--crash-reporter-id 4e28e8db-6680-4dea-a577-845468db680c|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|ubuntu|
|XDG_CURRENT_DESKTOP|Unity|
|XDG_SESSION_DESKTOP|ubuntu|
|XDG_SESSION_TYPE|x11|
</details><details>
<summary>Process Info</summary>
```
CPU % Mem MB PID Process
0 129 167128 code main
0 0 167132 zygote
0 129 167175 gpu-process
0 0 167210 broker
0 0 167133 zygote
0 0 167135 zygote
0 257 167191 window [1] (train.py - ECG_Encoder_for_JoLT - Visual Studio Code)
0 0 167177 utility-network-service
0 0 167318 ptyHost
0 0 167434 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh
0 0 167437 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh
0 0 167914 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh
0 0 168130 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh
0 386 167333 extensionHost [1]
0 0 167632 /home/revlis_ai/.vscode/extensions/ms-python.python-2024.22.2-linux-x64/python-env-tools/bin/pet server
0 0 168094 /usr/share/code/code /usr/share/code/resources/app/extensions/json-language-features/server/dist/node/jsonServerMain --node-ipc --clientProcessId=167333
0 257 179521 electron-nodejs (bundle.js )
0 129 167364 shared-process
0 0 180812 /bin/sh -c /usr/bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command=
0 0 180813 /usr/bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command=
0 129 167366 fileWatcher [1]
```
</details>
<details>
<summary>Workspace Info</summary>
```
| Window (train.py - ECG_Encoder_for_JoLT - Visual Studio Code)
| Folder (ECG_Encoder_for_JoLT): more than 44316 files
| File types: dat(9641) hea(9641) png(610) npz(19) hdf5(18) txt(8) pt(8)
| py(8) pyc(3) csv(3)
| Conf files:;
```
</details>
<details><summary>Extensions (23)</summary>
Extension|Author (truncated)|Version
---|---|---
arepl|alm|3.0.0
githistory|don|0.6.20
gitlens|eam|16.2.0
vscode-html-css|ecm|2.0.12
copilot|Git|1.257.0
copilot-chat|Git|0.23.2
vscode-docker|ms-|1.29.3
vscode-language-pack-zh-hant|MS-|1.96.2024121109
debugpy|ms-|2024.14.0
python|ms-|2024.22.2
vscode-pylance|ms-|2024.12.1
datawrangler|ms-|1.16.0
jupyter|ms-|2024.11.0
jupyter-keymap|ms-|1.1.2
jupyter-renderers|ms-|1.0.21
vscode-jupyter-cell-tags|ms-|0.1.9
vscode-jupyter-slideshow|ms-|0.1.6
remote-containers|ms-|0.394.0
remote-wsl|ms-|0.88.5
vscode-yaml|red|1.15.0
pdf|tom|1.2.2
vscode-pets|ton|1.27.0
tconda|zhp|0.0.5
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vspor879:30202332
vspor708:30202333
vspor363:30204092
vswsl492cf:30256860
vscod805cf:30301675
binariesv615:30325510
vsaa593cf:30376535
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
a9j8j154:30646983
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
h48ei257:31000450
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupyter:31046869
2f103344:31071589
nativerepl2:31139839
pythonrstrctxt:31112756
nativeloc1:31192215
cf971741:31144450
iacca1:31171482
notype1:31157159
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
6074i472:31201624
dwoutputs:31217127
```
</details>
<!-- generated by issue reporter --> | info-needed | low | Critical |
2,796,881,534 | tensorflow | Aborted in `tf.raw_ops.RaggedGather` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.17
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tf.raw_ops.RaggedGather` triggers crash.
### Standalone code to reproduce the issue
```shell
params_nested_splits = tf.constant(0, shape=[3], dtype=tf.int64)
params_dense_values = tf.constant(1, shape=[0], dtype=tf.float32)
indices = tf.constant(0, shape=[], dtype=tf.int64)
OUTPUT_RAGGED_RANK = 1
PARAMS_RAGGED_RANK = 1
tf.raw_ops.RaggedGather(
params_nested_splits=[params_nested_splits],
params_dense_values=params_dense_values,
indices=indices,
OUTPUT_RAGGED_RANK=1,
name=None
)
```
### Relevant log output
```shell
2025-01-18 09:30:00.549762: F tensorflow/core/framework/tensor.cc:844] Check failed: dtype() == expected_dtype (9 vs. 1) float expected, got int64
Aborted (core dumped)
``` | type:bug,comp:ops,2.17 | medium | Critical |
2,796,891,774 | flutter | 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 842 pos 12: '_hold == null': is not true. | ### Steps to reproduce
I wanted to create an instant return effect so that the header would always be open or closed.
Essentially this code works, but at some point it produces a critical error and scrolling stops working.
### Expected results
.
### Actual results
/
### Code sample
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
@override
Widget build(BuildContext context) {
return MaterialApp(
darkTheme: ThemeData.dark(),
themeMode: ThemeMode.dark,
);
}
}
class TelegramStyleAppBar extends StatefulWidget {
const TelegramStyleAppBar({super.key});
@override
_TelegramStyleAppBarState createState() => _TelegramStyleAppBarState();
}
class _TelegramStyleAppBarState extends State<TelegramStyleAppBar> {
late ScrollController _scrollController;
double expandedHeight = 300.0;
double collapsedHeight = 100.0;
@override
void initState() {
super.initState();
_scrollController = ScrollController();
// Инициализация слушателей
WidgetsBinding.instance.addPostFrameCallback(
(timeStamp) {
print(_scrollController.position);
_scrollController.position.isScrollingNotifier.addListener(
() {
if (!_scrollController.position.isScrollingNotifier.value) {
_handleSnapScroll();
}
},
);
},
);
}
@override
void dispose() {
_scrollController.dispose();
super.dispose();
}
@override
Widget build(BuildContext context) {
return Scaffold(
backgroundColor: Colors.black,
body: CustomScrollView(
controller: _scrollController,
physics: const BouncingScrollPhysics(),
slivers: [
SliverAppBar(
pinned: true,
stretch: true,
surfaceTintColor: Colors.transparent,
expandedHeight: expandedHeight,
collapsedHeight: collapsedHeight,
backgroundColor: Colors.black,
flexibleSpace: LayoutBuilder(
builder: (BuildContext context, BoxConstraints constraints) {
final appBarHeight = constraints.biggest.height;
final normalizedHeight = (appBarHeight - collapsedHeight) /
(expandedHeight - collapsedHeight);
final clampedHeight = normalizedHeight.clamp(0.0, 1.0);
return FlexibleSpaceBar(
centerTitle: true,
title: Opacity(
opacity: 1,
child: Text(
'Telegram Header',
style: TextStyle(
fontSize: 20 + clampedHeight * 10,
color: Colors.white,
),
),
),
background: Container(
color: Colors.blue,
child: Center(
child: Text(
'Pull to Stretch',
style: TextStyle(
fontSize: 16 + clampedHeight * 24,
color: Colors.white,
),
),
),
),
);
},
),
),
SliverList(
delegate: SliverChildBuilderDelegate(
(context, index) => ListTile(
title: Text('Item $index'),
),
childCount: 50,
),
),
],
),
);
}
void _handleSnapScroll() async {
double offset = _scrollController.offset;
double snapThreshold = (expandedHeight - collapsedHeight) / 2;
// Проверка, на какой позиции находится скролл
if (offset < snapThreshold) {
// Скролл до раскрытой шапки
await _scrollController.animateTo(
0.0,
duration: const Duration(milliseconds: 300),
curve: Curves.easeInOut,
);
} else if (offset < expandedHeight - collapsedHeight) {
// Скролл до свернутой шапки
await _scrollController.animateTo(
expandedHeight - collapsedHeight,
duration: const Duration(milliseconds: 300),
curve: Curves.easeInOut,
);
}
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
[Upload media here]
</details>
### Logs
<details open><summary>Logs</summary>
```console
Restarted application in 432ms.
flutter: ScrollPositionWithSingleContext#5d946(offset: 0.0, range: 0.0..2288.0, viewport: 874.0, ScrollableState, BouncingScrollPhysics -> BouncingScrollPhysics -> RangeMaintainingScrollPhysics, IdleScrollActivity#a82a1, ScrollDirection.idle)
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
[ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas.
════════ Exception caught by gesture ═══════════════════════════════════════════
The following assertion was thrown while handling a gesture:
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 842 pos 12: '_hold == null': is not true.
Either the assertion indicates an error in the framework itself, or we should provide substantially more information in this error message to help you determine and fix the underlying cause.
In either case, please report this assertion by filing a bug on GitHub:
https://github.com/flutter/flutter/issues/new?template=2_bug.yml
When the exception was thrown, this was the stack:
#2 ScrollableState._handleDragStart (package:flutter/src/widgets/scrollable.dart:842:12)
#3 DragGestureRecognizer._checkStart.<anonymous closure> (package:flutter/src/gestures/monodrag.dart:790:53)
#4 GestureRecognizer.invokeCallback (package:flutter/src/gestures/recognizer.dart:351:24)
#5 DragGestureRecognizer._checkStart (package:flutter/src/gestures/monodrag.dart:790:7)
#6 DragGestureRecognizer._checkDrag (package:flutter/src/gestures/monodrag.dart:757:5)
#7 DragGestureRecognizer.acceptGesture (package:flutter/src/gestures/monodrag.dart:684:7)
#8 GestureArenaManager._resolveByDefault (package:flutter/src/gestures/arena.dart:271:25)
#9 GestureArenaManager._tryToResolveArena.<anonymous closure> (package:flutter/src/gestures/arena.dart:251:31)
(elided 4 frames from class _AssertionError and dart:async)
Handler: "onStart"
Recognizer: VerticalDragGestureRecognizer#9732c
start behavior: start
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by gesture ═══════════════════════════════════════════
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 853 pos 12: '_hold == null || _drag == null': is not true.
════════════════════════════════════════════════════════════════════════════════
Reloaded 1 of 709 libraries in 394ms (compile: 38 ms, reload: 161 ms, reassemble: 135 ms).
Reloaded 1 of 709 libraries in 328ms (compile: 43 ms, reload: 103 ms, reassemble: 91 ms).
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
f%
a001@MacBook-exeshka telegram_ui % flutter doctor -v
[✓] Flutter (Channel stable, 3.24.5, on macOS 15.2 24C101 darwin-arm64, locale ru-KZ)
• Flutter version 3.24.5 on channel stable at /Users/a001/flutter
• Upstream repository https://github.com/flutter/flutter.git
• Framework revision dec2ee5c1f (9 weeks ago), 2024-11-13 11:13:06 -0800
• Engine revision a18df97ca5
• Dart version 3.5.4
• DevTools version 2.37.3
[✓] Android toolchain - develop for Android devices (Android SDK version 35.0.0)
• Android SDK at /Users/a001/Library/Android/sdk
• Platform android-35, build-tools 35.0.0
• Java binary at: /opt/homebrew/Cellar/openjdk@17/17.0.13/libexec/openjdk.jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment Homebrew (build 17.0.13+0)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 16.2)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Build 16C5032a
• CocoaPods version 1.16.2
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 2024.2)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 21.0.3+-79915917-b509.11)
[✓] VS Code (version 1.95.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.102.0
[✓] Connected device (5 available)
• iPhone exeshka (mobile) • 00008101-0015148A0C40001E • ios • iOS 18.2 22C150
• iPhone 16 Pro (mobile) • 54FA6352-39FF-44D9-8BB9-ADF8BCF9E69E • ios • com.apple.CoreSimulator.SimRuntime.iOS-18-2 (simulator)
• macOS (desktop) • macos • darwin-arm64 • macOS 15.2 24C101 darwin-arm64
• Mac Designed for iPad (desktop) • mac-designed-for-ipad • darwin • macOS 15.2 24C101 darwin-arm64
• Chrome (web) • chrome • web-javascript • Google Chrome 131.0.6778.266
[✓] Network resources
• All expected network resources are available.
• No issues found!
a001@MacBook-exeshka telegram_ui %
```
</details>
| waiting for customer response,in triage | low | Critical |
2,796,900,433 | vscode | Explorer not working as intended |
Type: <b>Bug</b>
The explorer doesnt show the folders and folder structure.
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Windows_NT x64 10.0.22631
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|11th Gen Intel(R) Core(TM) i5-11300H @ 3.10GHz (8 x 3110)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|undefined|
|Memory (System)|7.75GB (0.39GB free)|
|Process Argv|--crash-reporter-id f57f71a7-a455-4bb6-87be-c0f886d15903|
|Screen Reader|no|
|VM|0%|
</details><details><summary>Extensions (4)</summary>
Extension|Author (truncated)|Version
---|---|---
prettier-vscode|esb|11.0.0
copilot|Git|1.257.0
copilot-chat|Git|0.23.2
LiveServer|rit|5.7.9
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vspor879:30202332
vspor708:30202333
vspor363:30204092
vscod805:30301674
binariesv615:30325510
vsaa593cf:30376535
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
2e7ec940:31000449
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupytercf:31046870
nativerepl1:31139838
pythonrstrctxt:31112756
nativeloc1:31192215
cf971741:31144450
iacca1:31171482
notype1cf:31157160
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
3d9ag387:31215808
6074i472:31201624
dwoutputs:31217127
```
</details>
<!-- generated by issue reporter --> | info-needed | low | Critical |
2,796,906,847 | pytorch | Bracket indexing not working | ### 🐛 Describe the bug
Unsqueezing not working
```python
import torch
tn = torch.randn(6980, 1, 12, 16, 20)
tn[[1], :, :, :, :].shape # (1, 1, 12, 16, 20)
tn[[1], :, [11], :, :].shape # (1, 1, 16, 20)
```
### Versions
PyTorch version: N/A
Is debug build: N/A
CUDA used to build PyTorch: N/A
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: version 3.22.1
Libc version: glibc-2.35
Python version: 3.12.4 | packaged by Anaconda, Inc. | (main, Jun 18 2024, 15:12:24) [GCC 11.2.0] (64-bit runtime)
Python platform: Linux-6.8.0-51-generic-x86_64-with-glibc2.35
Is CUDA available: N/A
CUDA runtime version: 11.5.119
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration:
GPU 0: NVIDIA GeForce RTX 4090
GPU 1: NVIDIA GeForce RTX 4090
Nvidia driver version: 555.42.02
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: N/A
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 39 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 32
On-line CPU(s) list: 0-31
Vendor ID: GenuineIntel
Model name: 13th Gen Intel(R) Core(TM) i9-13900F
CPU family: 6
Model: 183
Thread(s) per core: 2
Core(s) per socket: 24
Socket(s): 1
Stepping: 1
CPU max MHz: 5600.0000
CPU min MHz: 800.0000
BogoMIPS: 3993.60
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb intel_pt sha_ni xsaveopt xsavec xgetbv1 xsaves split_lock_detect user_shstk avx_vnni dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp hwp_pkg_req hfi vnmi umip pku ospke waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm md_clear serialize arch_lbr ibt flush_l1d arch_capabilities
Virtualization: VT-x
L1d cache: 896 KiB (24 instances)
L1i cache: 1.3 MiB (24 instances)
L2 cache: 32 MiB (12 instances)
L3 cache: 36 MiB (1 instance)
NUMA node(s): 1
NUMA node0 CPU(s): 0-31
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Mitigation; Clear Register File
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI BHI_DIS_S
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] numpy==2.1.0
[conda] blas 1.0 mkl
[conda] cuda-cudart 12.1.105 0 nvidia
[conda] cuda-cupti 12.1.105 0 nvidia
[conda] cuda-libraries 12.1.0 0 nvidia
[conda] cuda-nvrtc 12.1.105 0 nvidia
[conda] cuda-nvtx 12.1.105 0 nvidia
[conda] cuda-opencl 12.3.52 0 nvidia
[conda] cuda-runtime 12.1.0 0 nvidia
[conda] easy-torch 1.3.2 pypi_0 pypi
[conda] ffmpeg 4.3 hf484d3e_0 pytorch
[conda] libcublas 12.1.0.26 0 nvidia
[conda] libcufft 11.0.2.4 0 nvidia
[conda] libcurand 10.3.4.52 0 nvidia
[conda] libcusolver 11.4.4.55 0 nvidia
[conda] libcusparse 12.0.2.55 0 nvidia
[conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch
[conda] libnvjitlink 12.1.105 0 nvidia
[conda] mkl 2023.1.0 h213fc3f_46343
[conda] mkl-service 2.4.0 py311h5eee18b_1
[conda] mkl_fft 1.3.8 py311h5eee18b_0
[conda] mkl_random 1.2.4 py311hdb19cb5_0
[conda] numpy 1.24.4 pypi_0 pypi
[conda] nvidia-cublas-cu12 12.1.3.1 pypi_0 pypi
[conda] nvidia-cuda-cupti-cu12 12.1.105 pypi_0 pypi
[conda] nvidia-cuda-nvrtc-cu12 12.1.105 pypi_0 pypi
[conda] nvidia-cuda-runtime-cu12 12.1.105 pypi_0 pypi
[conda] nvidia-cudnn-cu12 8.9.2.26 pypi_0 pypi
[conda] nvidia-cufft-cu12 11.0.2.54 pypi_0 pypi
[conda] nvidia-curand-cu12 10.3.2.106 pypi_0 pypi
[conda] nvidia-cusolver-cu12 11.4.5.107 pypi_0 pypi
[conda] nvidia-cusparse-cu12 12.1.0.106 pypi_0 pypi
[conda] nvidia-nccl-cu12 2.20.5 pypi_0 pypi
[conda] nvidia-nvjitlink-cu12 12.4.127 pypi_0 pypi
[conda] nvidia-nvtx-cu12 12.1.105 pypi_0 pypi
[conda] pytorch-cuda 12.1 ha16c6d3_5 pytorch
[conda] pytorch-forecasting 1.2.0 pypi_0 pypi
[conda] pytorch-lightning 2.2.0 pypi_0 pypi
[conda] pytorch-mutex 1.0 cuda pytorch
[conda] torch 2.3.0 pypi_0 pypi
[conda] torch-cluster 1.6.3+pt23cu121 pypi_0 pypi
[conda] torch-geometric 2.4.0 pypi_0 pypi
[conda] torch-scatter 2.1.2+pt23cu121 pypi_0 pypi
[conda] torch-sparse 0.6.18+pt23cu121 pypi_0 pypi
[conda] torch-spline-conv 1.2.2+pt23cu121 pypi_0 pypi
[conda] torch-summary 1.4.5 pypi_0 pypi
[conda] torchaudio 2.3.0 pypi_0 pypi
[conda] torchinfo 1.8.0 pypi_0 pypi
[conda] torchmetrics 1.3.0.post0 pypi_0 pypi
[conda] torchsummary 1.5.1 pypi_0 pypi
[conda] torchvision 0.18.0 pypi_0 pypi
[conda] triton 2.3.0 pypi_0 pypi | needs reproduction,triaged,module: advanced indexing | low | Critical |
2,796,928,799 | pytorch | Please add fp16 to MPS devices. | ### 🚀 The feature, motivation and pitch
I used torch==2.7 to train llama via huggingface transformers,but
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 9.00it/s]
trainable params: 4,194,304 || all params: 6,933,450,752 || trainable%: 0.0605
Traceback (most recent call last):
File "/Users/rbhan/Data/StellAIHub/Train/LanguageModel/xmt.py", line 115, in <module>
trainer = Trainer(
^^^^^^^^
File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/utils/deprecation.py", line 165, in wrapped_func
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/trainer.py", line 459, in __init__
self.create_accelerator_and_postprocess()
File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/trainer.py", line 5071, in create_accelerator_and_postprocess
self.accelerator = Accelerator(**args)
^^^^^^^^^^^^^^^^^^^
File "/Users/rbhan/Data/AIHub/Trans-Penv/accelerate/src/accelerate/accelerator.py", line 495, in __init__
raise ValueError(f"fp16 mixed precision requires a GPU (not {self.device.type!r}).")
ValueError: fp16 mixed precision requires a GPU (not 'mps').
### Alternatives
So when can pytorch support fp16 on MPS?I have waited for one year,but no solution.Thank you!
### Additional context
_No response_
cc @mcarilli @ptrblck @leslie-fang-intel @jgong5 @kulinseth @albanD @malfet @DenisVieriu97 @jhavukainen | needs reproduction,triaged,module: amp (automated mixed precision),module: mps | low | Critical |
2,796,933,056 | tensorflow | Aborted in ` tf.nn.conv3d_transpose` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
No
### Source
source
### TensorFlow version
tf 2.16.1
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tf.nn.conv3d_transpose` triggers crash.
### Standalone code to reproduce the issue
```shell
https://colab.research.google.com/drive/1wA5pUfm4HpNpHx-PawrZt4CKS2YCUR-4?usp=sharing
```
### Relevant log output
```shell
2025-01-18 11:31:33.696896: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2754] Non-OK-status: GetNodeAttr(orig_node->def(), "strides", &strides) status: INVALID_ARGUMENT: Attr strides has value 5717580618211388939 out of range for an int32
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:apis,comp:ops,TF 2.16 | medium | Critical |
2,796,938,285 | godot | project manager unable to resolve icon referenced by uid before import | ### Tested versions
4.4-beta1
### System information
(Native)
Godot v4.4.beta1 - Windows 10 (build 19045) - Multi-window, 2 monitors - OpenGL 3 (Compatibility) - NVIDIA GeForce RTX 4070 (NVIDIA; 32.0.15.6636) - AMD Ryzen 9 3950X 16-Core Processor (32 threads)
(WSL2 on above system)
Godot v4.4.beta1 - Ubuntu 24.04.1 LTS 24.04 - X11 display driver, Multi-window, 2 monitors - OpenGL 3 (Compatibility) - D3D12 (NVIDIA GeForce RTX 4070) - AMD Ryzen 9 3950X 16-Core Processor (32 threads)
### Issue description
The project manager is unable to resolve an icon referenced by uid before first import:
```
ERROR: Unrecognized UID: "uid://b3xkos3u4jns4".
at: get_id_path (core/io/resource_uid.cpp:157)
ERROR: Error opening file ''.
at: load_image (core/io/image_loader.cpp:89)
```
### Steps to reproduce
open project manager
import MRP without "Edit Now"
-> no icon, errors in console
### Minimal reproduction project (MRP)
[icontest44b1.zip](https://github.com/user-attachments/files/18464271/icontest44b1.zip) | bug,topic:editor | low | Critical |
2,796,939,108 | godot | GPUParticles3D collide with their gizmo icon, which can make them appear to not move at all if using the Point emission mode | ### Tested versions
- Reproducible in: 4.4.beta1, 4.4.beta 7b1ed520b
- Not reproducible in: 4.3.stable
### System information
Godot v4.4.beta (7b1ed520b) - Fedora Linux 41 (KDE Plasma) on X11 - X11 display driver, Multi-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 4090 (nvidia; 565.77) - 13th Gen Intel(R) Core(TM) i9-13900K (32 threads)
### Issue description
GPUParticles3D collide with their gizmo icon, which can make them appear to not move at all if using the Point emission mode.
I bisected the regression to https://github.com/godotengine/godot/pull/91174. This likely occurs because making the gizmo material use alpha scissor means it's now taken into account by the heightfield collision generator. We can probably fix this by changing the cull mask of the GPUParticlesCollisionHeightField3D to exclude the editor-only gizmo layers (`Node3DEditorViewport::GIZMO_EDIT_LAYER`).
This is further proven by the collision's shape matching the GPUParticles3D gizmo icon. You can see this by using a sphere emission shape and greatly increasing the particle count:

Relevant code:
https://github.com/godotengine/godot/blob/7b1ed520bdcca21d36e25e2094802b9b33dae2c6/editor/plugins/node_3d_editor_gizmos.cpp#L831-L837
https://github.com/godotengine/godot/blob/7b1ed520bdcca21d36e25e2094802b9b33dae2c6/scene/3d/gpu_particles_collision_3d.cpp
### Steps to reproduce
- Set up a scene with various meshes and particles that have collision enabled.
- Add a GPUParticlesCollisionHeightField3D node.
- Save the scene and use **Scene > Reload Saved Scene**.
- Notice how particles no longer move due to colliding with their gizmo.
### Minimal reproduction project (MRP)
[test_particle_collision_gizmo.zip](https://github.com/user-attachments/files/18464280/test_particle_collision_gizmo.zip) | bug,topic:editor,regression,topic:3d,topic:particles | low | Minor |
2,796,940,106 | pytorch | [RFC] Improve performance for layer_norm op for cuda with revectorized | ### 🚀 The feature, motivation and pitch
I found there is big perf drop if the size of layer_norm's inside size is not multiple of 4, there is a micro test case:
```python
import torch
DEVICE=torch.device('cuda')
# Time cost for near 1024
for cnt in range(2040, 2050):
x = torch.randn(4096, cnt, device=DEVICE, dtype=torch.float32)
w_shape = (x.shape[-1], )
#warm up
need_warmup = True
round = 5
if need_warmup:
for _ in range(round):
output = torch.nn.functional.layer_norm(x, w_shape)
torch.cuda.synchronize()
start_time = torch.cuda.Event(enable_timing=True)
end_time = torch.cuda.Event(enable_timing=True)
# Start time
start_time.record()
# Apply layernorm
for _ in range(round):
output = torch.nn.functional.layer_norm(x, w_shape)
# End time
end_time.record()
torch.cuda.synchronize()
# Calculate elapsed time
elapsed_time_ms = start_time.elapsed_time(end_time)
# print(f"CUDA Time: {elapsed_time_ms:.6f} ms")
gbps = lambda ms: round * 2 * x.numel() * x.element_size() * 1e-9 / (ms * 1e-3)
print(f"n as {cnt} of softmax: {gbps(elapsed_time_ms):.6f} gb/s")
```
Its output is:
```
n as 2040 of softmax: 483.555543 gb/s
n as 2041 of softmax: 345.531858 gb/s
n as 2042 of softmax: 345.369984 gb/s
n as 2043 of softmax: 347.825623 gb/s
n as 2044 of softmax: 489.580822 gb/s
n as 2045 of softmax: 345.591973 gb/s
n as 2046 of softmax: 346.057951 gb/s
n as 2047 of softmax: 345.850055 gb/s
n as 2048 of softmax: 470.192376 gb/s
n as 2049 of softmax: 347.012446 gb/s
```
We could see for the perf for input with `N = 2040, 2044, 2048` is obvious greater than other input `(480 vs 340)`.
Therefore, what I would like to do is **mitigating the perf gap between these input**.
### Alternatives
The root cause is that there is two kernels for `layer_norm`:
https://github.com/pytorch/pytorch/blob/5e4cf3e6ad6f1f06436f409b394ae02e5ed5583d/aten/src/ATen/native/cuda/layer_norm_kernel.cu#L824-L837
We could see if `N` is multiple of `num_vec_elems (4)`, it will call a kernel called `launch_vectorized_layer_norm_kernel`, which could load vectorized elements.
So what we could do is to also enable vectorized elements load for those case whose `N` is not multiple for `num_vec_elems (4)`.
I tried a draft implement to achieve it, it could improve performance same as the vectorized case, for this case it's about **~40%**, `(480 vs 340) take 2042 as example`.
Optimized data:
```
n as 2040 of softmax: 459.758796 gb/s
n as 2041 of softmax: 497.804886 gb/s
n as 2042 of softmax: 479.061575 gb/s
n as 2043 of softmax: 477.197096 gb/s
n as 2044 of softmax: 473.285078 gb/s
n as 2045 of softmax: 473.795206 gb/s
n as 2046 of softmax: 495.999985 gb/s
n as 2047 of softmax: 499.649134 gb/s
n as 2048 of softmax: 455.111104 gb/s
n as 2049 of softmax: 473.928423 gb/s
```
### Additional context
Here is a [doc](https://docs.google.com/document/d/1TFIbJAO3tek1-EltVvMC_0TgYuEeBZw8rxDrSb_Wx8Q/edit?tab=t.0) contains some stuffs about it.
And there is another optimization, we could see there is still perf gap for `layer_norm` in `pytorch` and `triton`, we may mitigate the gap by using **register to cache the data in pytorch kernel**, because the `launch_vectorized_layer_norm_kernel` load data from gmem twice yet, but I guess the `N` for `layer_norm` op may usually be a very big number (> 10k for some 2d layer_norm), so it may introduce much register pressure, WDYT?
cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @ptrblck @msaroufim @eqy | module: nn,module: cuda,triaged,topic: performance | low | Major |
2,796,945,049 | tensorflow | Abort in `tensorflow.keras.backend.conv2d` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
No
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.keras.backend.conv2d` triggers crash.
The crash occurs in the operator Conv2D
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
x = np.array([1.5841993e+38, 4.7480855e+36, -1.8916006e+38], dtype=np.float32)
kernel = np.ones((13, 8, 5, 14), dtype=np.float32)
tensorflow.keras.backend.conv2d(x=x, kernel=kernel)
```
### Relevant log output
```shell
2025-01-18 12:02:15.375499: F ./tensorflow/core/util/tensor_format.h:428] Check failed: index >= 0 && index < num_total_dims Invalid index from the dimension: 3, 0, C
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:apis,2.17 | medium | Critical |
2,796,947,730 | tensorflow | Aborted (core dumped) in `tensorflow.nn.max_pool1d` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.nn.max_pool1d` triggers crash.
The crash occurs in the operator MaxPool
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = np.ones((17,7,14), dtype=np.float32)
ksize = [3134755050008138495]
strides = [48732]
padding = 'SAME'
tensorflow.nn.max_pool1d(input=input, ksize=ksize, strides=strides, padding=padding)
```
### Relevant log output
```shell
2025-01-18 12:09:37.053592: F tensorflow/core/common_runtime/mkl_layout_pass.cc:1609] Non-OK-status: GetNodeAttr(n->def(), "ksize", &ksize) status: INVALID_ARGUMENT: Attr ksize has value 3134755050008138495 out of range for an int32
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,949,006 | tensorflow | Floating point exception in `tensorflow.nn.max_pool1d` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.nn.max_pool1d` triggers crash.
The crash occurs in the operator MaxPool
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = np.ones((16,19,18), dtype=np.float32)
ksize = [58486]
strides = [49657]
padding = 'VALID'
tensorflow.nn.max_pool1d(input=input, ksize=ksize, strides=strides, padding=padding)
```
### Relevant log output
```shell
Floating point exception (core dumped)
``` | stat:awaiting response,type:bug,comp:apis,comp:ops,2.17 | medium | Critical |
2,796,950,022 | tensorflow | Aborted in `tensorflow.nn.max_pool3d` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.nn.max_pool3d` triggers crash.
The crash occurs in the operator MaxPool3D
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = np.ones((20, 19, 20, 5, 0), dtype=np.float32)
ksize = [3505357736]
strides = [2230364274603370470]
padding = 'SAME'
tensorflow.nn.max_pool3d(input=input, ksize=ksize, strides=strides, padding=padding)
```
### Relevant log output
```shell
2025-01-18 12:15:49.067137: F tensorflow/core/common_runtime/mkl_layout_pass.cc:1609] Non-OK-status: GetNodeAttr(n->def(), "ksize", &ksize) status: INVALID_ARGUMENT: Attr ksize has value 3505357736 out of range for an int32
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,951,107 | rust | Tracking Issue for `stdarch_s390x` | Feature gate: `#![feature(stdarch_s390x)]`
This is a tracking issue for the s390x (aka SystemZ) intrinsics in `core::arch::s390x`.
### Public API
Everything in `core::arch::s390x`.
Missing instructions
based on the GCC [vecintrin.h](https://github.com/gcc-mirror/gcc/blob/master/gcc/config/s390/vecintrin.h)
- [ ] `vec_step`
- [ ] `__lcbb`
- [ ] `vec_adde`
- [ ] `vec_addec`
- [ ] `vec_all_nle`
- [ ] `vec_all_nlt`
- [ ] `vec_any_nle`
- [ ] `vec_any_nlt`
- [ ] `vec_genmask`
- [ ] `vec_genmasks_8`
- [ ] `vec_genmasks_16`
- [ ] `vec_genmasks_32`
- [ ] `vec_genmasks_64`
- [ ] `vec_splat_u8`
- [ ] `vec_splat_s8`
- [ ] `vec_splat_u16`
- [ ] `vec_splat_s16`
- [ ] `vec_splat_u32`
- [ ] `vec_splat_s32`
- [ ] `vec_splat_u64`
- [ ] `vec_splat_s64`
- [ ] `vec_sube`
- [ ] `vec_subec`
- [ ] `vec_checksum`
- [ ] `vec_gfmsum_128`
- [ ] `vec_gfmsum_accum_128`
- [ ] `vec_ceil`
- [ ] `vec_roundp`
- [ ] `vec_floor`
- [ ] `vec_roundm`
- [ ] `vec_trunc`
- [ ] `vec_roundz`
- [ ] `vec_rint`
- [ ] `vec_roundc`
- [ ] `vec_round`
- [ ] `vec_doublee`
- [ ] `vec_floate`
- [ ] `vec_all_nan`
- [ ] `vec_all_numeric`
- [ ] `vec_any_nan`
- [ ] `vec_any_numeric`
- [ ] `vec_extend_to_fp32_hi`
- [ ] `vec_extend_to_fp32_lo`
- [ ] `vec_round_from_fp32`
- [ ] `vec_convert_to_fp16`
- [ ] `vec_convert_from_fp16`
- [ ] `vec_gen_element_masks_8`
- [ ] `vec_gen_element_masks_16`
- [ ] `vec_gen_element_masks_32`
- [ ] `vec_gen_element_masks_64`
- [ ] `vec_gen_element_masks_128`
- [x] `vec_abs`
- [ ] `vec_add_u128`
- [ ] `vec_addc`
- [ ] `vec_addc_u128`
- [ ] `vec_adde_u128`
- [ ] `vec_addec_u128`
- [ ] `vec_all_eq`
- [ ] `vec_all_ge`
- [ ] `vec_all_gt`
- [ ] `vec_all_le`
- [ ] `vec_all_lt`
- [ ] `vec_all_ne`
- [ ] `vec_all_nge`
- [ ] `vec_all_ngt`
- [ ] `vec_and`
- [ ] `vec_andc`
- [ ] `vec_any_eq`
- [ ] `vec_any_ge`
- [ ] `vec_any_gt`
- [ ] `vec_any_le`
- [ ] `vec_any_lt`
- [ ] `vec_any_ne`
- [ ] `vec_any_nge`
- [ ] `vec_any_ngt`
- [ ] `vec_avg`
- [ ] `vec_blend`
- [ ] `vec_bperm`
- [ ] `vec_bperm_u128`
- [ ] `vec_cmpeq`
- [ ] `vec_cmpeq_idx`
- [ ] `vec_cmpeq_idx_cc`
- [ ] `vec_cmpeq_or_0_idx`
- [ ] `vec_cmpeq_or_0_idx_cc`
- [ ] `vec_cmpge`
- [ ] `vec_cmpgt`
- [ ] `vec_cmple`
- [ ] `vec_cmplt`
- [ ] `vec_cmpne_idx`
- [ ] `vec_cmpne_idx_cc`
- [ ] `vec_cmpne_or_0_idx`
- [ ] `vec_cmpne_or_0_idx_cc`
- [ ] `vec_cmpnrg`
- [ ] `vec_cmpnrg_cc`
- [ ] `vec_cmpnrg_idx`
- [ ] `vec_cmpnrg_idx_cc`
- [ ] `vec_cmpnrg_or_0_idx`
- [ ] `vec_cmpnrg_or_0_idx_cc`
- [ ] `vec_cmprg`
- [ ] `vec_cmprg_cc`
- [ ] `vec_cmprg_idx`
- [ ] `vec_cmprg_idx_cc`
- [ ] `vec_cmprg_or_0_idx`
- [ ] `vec_cmprg_or_0_idx_cc`
- [ ] `vec_cntlz`
- [ ] `vec_cnttz`
- [ ] `vec_cp_until_zero`
- [ ] `vec_cp_until_zero_cc`
- [ ] `vec_ctd`
- [ ] `vec_ctd_s64`
- [ ] `vec_ctd_u64`
- [ ] `vec_ctsl`
- [ ] `vec_ctul`
- [ ] `vec_double`
- [ ] `vec_eqv`
- [ ] `vec_evaluate`
- [ ] `vec_extend_s64`
- [ ] `vec_extract`
- [ ] `vec_find_any_eq`
- [ ] `vec_find_any_eq_cc`
- [ ] `vec_find_any_eq_idx`
- [ ] `vec_find_any_eq_idx_cc`
- [ ] `vec_find_any_eq_or_0_idx`
- [ ] `vec_find_any_eq_or_0_idx_cc`
- [ ] `vec_find_any_ne`
- [ ] `vec_find_any_ne_cc`
- [ ] `vec_find_any_ne_idx`
- [ ] `vec_find_any_ne_idx_cc`
- [ ] `vec_find_any_ne_or_0_idx`
- [ ] `vec_find_any_ne_or_0_idx_cc`
- [ ] `vec_float`
- [ ] `vec_fp_test_data_class`
- [ ] `vec_gather_element`
- [ ] `vec_gfmsum`
- [ ] `vec_gfmsum_accum`
- [ ] `vec_insert`
- [ ] `vec_insert_and_zero`
- [ ] `vec_ld2f`
- [ ] `vec_load_bndry`
- [ ] `vec_load_len`
- [ ] `vec_load_len_r`
- [ ] `vec_load_pair`
- [ ] `vec_madd`
- [x] `vec_max`
- [ ] `vec_meadd`
- [ ] `vec_mergeh`
- [ ] `vec_mergel`
- [ ] `vec_mhadd`
- [x] `vec_min`
- [ ] `vec_mladd`
- [ ] `vec_moadd`
- [ ] `vec_msub`
- [ ] `vec_msum`
- [ ] `vec_msum_u128`
- [ ] `vec_mule`
- [ ] `vec_mulh`
- [ ] `vec_mulo`
- [ ] `vec_nabs`
- [ ] `vec_nand`
- [ ] `vec_nmadd`
- [ ] `vec_nmsub`
- [ ] `vec_nor`
- [ ] `vec_or`
- [ ] `vec_orc`
- [ ] `vec_pack`
- [ ] `vec_packs`
- [ ] `vec_packs_cc`
- [ ] `vec_packsu`
- [ ] `vec_packsu_cc`
- [ ] `vec_perm`
- [ ] `vec_permi`
- [ ] `vec_popcnt`
- [ ] `vec_promote`
- [ ] `vec_revb`
- [ ] `vec_reve`
- [ ] `vec_rl`
- [ ] `vec_rl_mask`
- [ ] `vec_rli`
- [ ] `vec_scatter_element`
- [ ] `vec_search_string_cc`
- [ ] `vec_search_string_until_zero_cc`
- [ ] `vec_sel`
- [ ] `vec_signed`
- [ ] `vec_slb`
- [ ] `vec_sld`
- [ ] `vec_sldb`
- [ ] `vec_sldw`
- [ ] `vec_sll`
- [ ] `vec_splat`
- [x] `vec_splats`
- [ ] `vec_sqrt`
- [ ] `vec_srab`
- [ ] `vec_sral`
- [ ] `vec_srb`
- [ ] `vec_srdb`
- [ ] `vec_srl`
- [ ] `vec_st2f`
- [ ] `vec_store_len`
- [ ] `vec_store_len_r`
- [ ] `vec_sub_u128`
- [ ] `vec_subc`
- [ ] `vec_subc_u128`
- [ ] `vec_sube_u128`
- [ ] `vec_subec_u128`
- [ ] `vec_sum`
- [ ] `vec_sum2`
- [ ] `vec_sum4`
- [ ] `vec_sum_u128`
- [ ] `vec_test_mask`
- [ ] `vec_unpackh`
- [ ] `vec_unpackl`
- [ ] `vec_unsigned`
- [ ] `vec_xl`
- [ ] `vec_xld2`
- [ ] `vec_xlw4`
- [ ] `vec_xor`
- [ ] `vec_xst`
- [ ] `vec_xstd2`
- [ ] `vec_xstw4`
### Steps / History
<!--
For larger features, more steps might be involved.
If the feature is changed later, please add those PRs here as well.
-->
- [ ] Implementation: #...
- [ ] Final comment period (FCP)[^1]
- [ ] Stabilization PR
<!--
Once the feature has gone through a few release cycles and there are no
unresolved questions left, the feature might be ready for stabilization.
If this feature didn't go through the RFC process, a final comment period
(FCP) is always needed before stabilization. This works as follows:
A library API team member can kick off the stabilization process, at which point
the rfcbot will ask all the team members to verify they agree with
stabilization. Once enough members agree and there are no concerns, the final
comment period begins: this issue will be marked as such and will be listed
in the next This Week in Rust newsletter. If no blocking concerns are raised in
that period of 10 days, a stabilization PR can be opened by anyone.
-->
### Unresolved Questions
<!--
Include any open questions that need to be answered before the feature can be
stabilised. If multiple (unrelated) big questions come up, it can be a good idea
to open a separate issue for each, to make it easier to keep track of the
discussions.
It's useful to link any relevant discussions and conclusions (whether on GitHub,
Zulip, or the internals forum) here.
-->
- None yet.
[^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html
@rustbot label O-SystemZ
general s390x vector/intrinsics progress is tracked at https://github.com/rust-lang/rust/issues/130869
cc @taiki-e | T-libs-api,C-tracking-issue,O-SystemZ | low | Minor |
2,796,952,009 | tensorflow | Aborted in `tensorflow.raw_ops.RecordInput` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.raw_ops.RecordInput` triggers crash.
The crash occurs in the operator RecordInput
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
file_pattern = '@\x0chnFvGe#83FxIIw.\';-$`E?8Y<>-C2p>\x0bl\\B*;z;]a4lrH{gWK}rAY{;T]2k&j-EB)pQu{JLy1K{Q|;g:>CEg<hA"^((UXJ?fMf\'NsZ@KK\rjp{mfE \tYx-3H3;5z\t(%g>\r<!G&_jqx-@o\rz"1 NS"eiM/ 6X/r+/\rj8h>6k{-B]P\n>[\'s0FOZ\x0b[aV\x0c1q\\!{D~2ZIBe{8v9;,oG=Hjsz&~.$T\x0b[_%/ 2;E:Hy8H\tx|dUPg2;[c_\'(>0z-v[Yob\n\x0c;aKX*F.h E@>\x0bK| zGnevJS7G=N_{c}\x0cS4p\']#1`T?FAep'
file_random_seed = 0
batch_size = -8494458287779632258
tensorflow.raw_ops.RecordInput(file_pattern=file_pattern, file_random_seed=file_random_seed, batch_size=batch_size)
```
### Relevant log output
```shell
2025-01-18 12:21:05.765578: F tensorflow/core/framework/tensor_shape.cc:201] Non-OK-status: InitDims(dim_sizes) status: INVALID_ARGUMENT: Expected shape dimensions to be non-negative, got -8494458287779632258
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,953,954 | tensorflow | Aborted in `tensorflow.compat.v1.nn.depthwise_conv2d_native` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.compat.v1.nn.depthwise_conv2d_native` triggers crash.
The crash occurs in the operator DepthwiseConv2DNative
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = -18261.482131447112
strides = [13762484222233278712, 17895496137899471100]
padding = 'VALID'
dilations = [11529923499597944069, 6104546329631715214, 16719022861480294065, 15641088296656225516, 5831796171324899039, 2019887057379389965, 6603155403475823832, 128243198633280011, 1962111378253501566, 12787126510677041415, 10754291262680183180]
filter = np.array([[ 36832. , -8776. , 255.2, -31200. , -6408. , -12656. , -31696. , -65024. , -50688. ]], dtype=np.float16)
tensorflow.compat.v1.nn.depthwise_conv2d_native(input=input, filter=filter, strides=strides, padding=padding, dilations=dilations)
```
### Relevant log output
```shell
2025-01-18 12:25:34.220155: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2755] Non-OK-status: GetNodeAttr(orig_node->def(), "dilations", &dilations) status: INVALID_ARGUMENT: Attr dilations has value 6104546329631715214 out of range for an int32
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,954,403 | electron | Mouse events in unfocused window | ### Preflight Checklist
- [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project.
- [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success.
### Electron Version
34.0.0
### What operating system(s) are you using?
macOS
### Operating System Version
15.2
### What arch are you using?
arm64 (including Apple Silicon)
### Last Known Working Electron version
_No response_
### Expected Behavior
`mouseleave` `mouseout` The timing of these should be the same whether the window has focus or not.
### Actual Behavior
When the mouse cursor moves over or off a focused window, it works normally.
However, when you move the mouse cursor over an unfocused window, `mouseleave` and `mouseout` occur immediately after `mouseenter` and `mouseover`. And when you move the mouse cursor off, `mouseleave` and `mouseout` do not occur.
My environment:
- Electron v34.0.0
- macOS 15.2 arm64
### Testcase Gist URL
_No response_
### Additional Information
I created a minimal application to recreate this.
I also created a video of the reenactment.
https://github.com/sdr0x07b6/electron-mouse-events-test
https://github.com/user-attachments/assets/bc13d84f-edb0-4bdc-9a05-c7200edfb531 | platform/macOS,bug :beetle:,has-repro-repo,34-x-y | low | Critical |
2,796,955,278 | tensorflow | Aborted in `tensorflow.raw_ops.ScatterNdNonAliasingAdd` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.raw_ops.ScatterNdNonAliasingAdd` triggers crash.
The crash occurs in the operator ScatterNdNonAliasingAdd
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = np.array([False, True, True, False, True, False, True, False, False,
False, False, True, True, False, False, True, True])
indices = 1547981256
updates = np.array([ True, False, True, False, True])
tensorflow.raw_ops.ScatterNdNonAliasingAdd(input=input, indices=indices, updates=updates)
```
### Relevant log output
```shell
2025-01-18 12:28:50.452498: F ./tensorflow/core/framework/tensor.h:847] Check failed: new_num_elements == NumElements() (1 vs. 5)
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,957,488 | tensorflow | Aborted in `tensorflow.compat.v1.nn.conv2d_backprop_filter` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.compat.v1.nn.conv2d_backprop_filter` triggers crash.
The crash occurs in the operator Conv2DBackpropFilter
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
input = np.array([1.2807603e+38], dtype=np.float32)
filter_sizes = np.ones((20, 6, 5, 3), dtype=np.uint16)
strides = [4538426903074820908]
padding = 'VALID'
out_backprop = np.ones((19, 5, 11, 4), dtype=np.float32)
tensorflow.compat.v1.nn.conv2d_backprop_filter(input=input, filter_sizes=filter_sizes, out_backprop=out_backprop, strides=strides, padding=padding)
```
### Relevant log output
```shell
2025-01-18 12:34:39.686592: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2754] Non-OK-status: GetNodeAttr(orig_node->def(), "strides", &strides) status: INVALID_ARGUMENT: Attr strides has value 4538426903074820908 out of range for an int32
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,958,241 | storybook | [Bug]: Next.js-Vite fails in new Next.js project | ### Describe the bug
#### What I did
- `pnpx create next-app`
- `pnpx storybook@latest init`
- manually replace `@storybook/nextjs` with `@storybook/experimental-nextjs-vite`

#### Workaround
- Manually install `styled-jsx`
### Reproduction link
-
### Reproduction steps
_No response_
### System
```bash
any?
```
### Additional context
_No response_ | bug,dependencies,has workaround,nextjs,sev:S2 | low | Critical |
2,796,960,526 | vscode | Git - Add autocomplete for SCM's commit message | Automatically complete commit message based on files in the staging.
I believe this is an exceptionally innovative and highly practical feature.
<img width="707" alt="Image" src="https://github.com/user-attachments/assets/9709cd24-9486-428d-b9ce-561f8fdf98a8" /> | feature-request,git | low | Minor |
2,796,962,703 | tensorflow | Aborted in `tensorflow.nn.depthwise_conv2d_backprop_filter` | ### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.16.1 tf 2.17.0
### Custom code
Yes
### OS platform and distribution
Ubuntu 20.04
### Mobile device
_No response_
### Python version
3.11
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
On specific inputs, `tensorflow.nn.depthwise_conv2d_backprop_filterr` triggers crash.
The crash occurs in the operator DepthwiseConv2dNativeBackpropFilter
### Standalone code to reproduce the issue
```shell
import pickle
import numpy as np
import tensorflow
filter_sizes = np.ones((13, 2, 5, 4), dtype=np.uint32)
input = 26262.2175547925
out_backprop = -14557.552005335412
strides = []
padding = 'VALID'
tensorflow.nn.depthwise_conv2d_backprop_filter(input=input, filter_sizes=filter_sizes, out_backprop=out_backprop, strides=strides, padding=padding)
```
### Relevant log output
```shell
2025-01-18 12:47:13.553183: F ./tensorflow/core/util/tensor_format.h:428] Check failed: index >= 0 && index < num_total_dims Invalid index from the dimension: 0, 0, N
Aborted (core dumped)
``` | stat:awaiting response,type:bug,comp:ops,2.17 | medium | Critical |
2,796,977,129 | yt-dlp | [Google Drive] Cannot download source file when cookies are given | ### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE
- [x] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field
### Checklist
- [x] I'm reporting that yt-dlp is broken on a **supported** site
- [x] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
- [x] I've checked that all provided URLs are playable in a browser with the same IP and same login details
- [x] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command)
- [x] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
- [x] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue)
- [x] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required
### Region
worldwide
### Provide a description that is worded well enough to be understood
When trying to download the source file for a google drive file, I seem to be getting Error 400 when cookies are provided. (example provided below). Now for publicly available files its not much of an issue to run without cookies to get the source file, but for privately shared files, or those on my own drive, I won't be able to access them without cookies, also video files that haven't been processed and don't have any other option other than source.
Not sure if the below are related, not knowledgeable enough to tell, but doesn't seem like the exact same issue.
[#4580](https://github.com/yt-dlp/yt-dlp/issues/4580)
[#9313](https://github.com/yt-dlp/yt-dlp/issues/9313)
If you are wondering why I would download files with yt-dlp from google drive, its is because I use MPV, which uses a yt-dlp hook, to stream google drive files.
### Provide verbose output that clearly demonstrates the problem
- [x] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`)
- [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead
- [x] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below
### Complete Verbose Output
```shell
*/ With cookies /*
yt-dlp -vU -F --cookies-from-browser=firefox https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing
[debug] Command-line config: ['-vU', '-F', '--cookies-from-browser=firefox', 'https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing']
[debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version [email protected] from yt-dlp/yt-dlp [c8541f8b1] (pip)
[debug] Python 3.13.1 (CPython AMD64 64bit) - Windows-10-10.0.19045-SP0 (OpenSSL 3.0.15 3 Sep 2024)
[debug] exe versions: ffmpeg 2023-08-28-git-b5273c619d-full_build-www.gyan.dev (setts), ffprobe 2023-08-28-git-b5273c619d-full_build-www.gyan.dev
[debug] Optional libraries: Cryptodome-3.21.0, brotli-1.1.0, certifi-2024.08.30, mutagen-1.47.0, requests-2.32.3, sqlite3-3.45.3, urllib3-2.2.3, websockets-14.1
[debug] Proxy map: {}
Extracting cookies from firefox
[debug] Extracting cookies from: "C:\Users\Name\AppData\Roaming\Mozilla\Firefox\Profiles\4on6caty.default\cookies.sqlite"
Extracted 2771 cookies from firefox
[debug] Request Handlers: urllib, requests, websockets
[debug] Loaded 1837 extractors
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
Latest version: [email protected] from yt-dlp/yt-dlp
yt-dlp is up to date ([email protected] from yt-dlp/yt-dlp)
[GoogleDrive] Extracting URL: https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing
[GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Downloading video webpage
[GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting source file
[GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting confirmed source file
WARNING: [GoogleDrive] Unable to request confirmed source file: HTTP Error 400: Bad Request
[debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, hdr:12(7), vcodec, channels, acodec, size, br, asr, proto, vext, aext, hasaud, source, id
[info] Available formats for 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-:
ID EXT RESOLUTION │ PROTO │ VCODEC ACODEC
───────────────────────────────────────────
18 mp4 640x360 │ https │ unknown unknown
22 mp4 1280x720 │ https │ unknown unknown
37 mp4 1920x1080 │ https │ unknown unknown
*/ without cookies /*
yt-dlp -vU -F https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing
[debug] Command-line config: ['-vU', '-F', 'https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing']
[debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version [email protected] from yt-dlp/yt-dlp [c8541f8b1] (pip)
[debug] Python 3.13.1 (CPython AMD64 64bit) - Windows-10-10.0.19045-SP0 (OpenSSL 3.0.15 3 Sep 2024)
[debug] exe versions: ffmpeg 2023-08-28-git-b5273c619d-full_build-www.gyan.dev (setts), ffprobe 2023-08-28-git-b5273c619d-full_build-www.gyan.dev
[debug] Optional libraries: Cryptodome-3.21.0, brotli-1.1.0, certifi-2024.08.30, mutagen-1.47.0, requests-2.32.3, sqlite3-3.45.3, urllib3-2.2.3, websockets-14.1
[debug] Proxy map: {}
[debug] Request Handlers: urllib, requests, websockets
[debug] Loaded 1837 extractors
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
Latest version: [email protected] from yt-dlp/yt-dlp
yt-dlp is up to date ([email protected] from yt-dlp/yt-dlp)
[GoogleDrive] Extracting URL: https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing
[GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Downloading video webpage
[GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting source file
[debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, hdr:12(7), vcodec, channels, acodec, size, br, asr, proto, vext, aext, hasaud, source, id
[info] Available formats for 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-:
ID EXT RESOLUTION │ PROTO │ VCODEC ACODEC
───────────────────────────────────────────────
18 mp4 640x360 │ https │ unknown unknown
22 mp4 1280x720 │ https │ unknown unknown
37 mp4 1920x1080 │ https │ unknown unknown
source mp4 unknown │ https │ unknown unknown
``` | account-needed,site-bug,triage,can-share-account | low | Critical |
2,796,983,161 | flutter | [go_router] App link provided to an app is ignored by GoRouter, when the app is suspended and resumed from background (Android) | ### Steps to reproduce
Hello,
I have an app with native GoRouter restoration implemented (by providing restorationScopeId).
When I provide an app link to my app, which is in the background and suspended by the OS to save memory, GoRouter ignores the app link and restores the route from its previous state. However, when the app is not started, in the foreground, or not suspended, the app link works as intended. GoRouter successfully opens the route provided by the link. Can you please advise on the issue? Thank you!
### Expected results
GoRouter opens the route according to the app link when an app link is provided to an app that is suspended in the background.
### Actual results
When an app link is provided to an app that is suspended in the background, GoRouter ignores the app link and restores the last route.
### Code sample
I don't think that providing a code sample is relevant here
### Screenshots or Video
### Logs
### Flutter Doctor output
flutter doctor -v
[✓] Flutter (Channel stable, 3.27.2, on Linux Mint 21.3 5.15.0-130-generic, locale en_US.UTF-8)
• Flutter version 3.27.2 on channel stable at /home/ruslan/fvm/versions/3.27.2
• Upstream repository ssh://[email protected]/flutter/flutter.git
• Framework revision 68415ad1d9 (5 days ago), 2025-01-13 10:22:03 -0800
• Engine revision e672b006cb
• Dart version 3.6.1
• DevTools version 2.40.2 | platform-android,package,has reproducible steps,p: go_router,team-go_router,found in release: 3.27,found in release: 3.28 | low | Major |
2,796,983,732 | ollama | Deepseek-v3 make smaller quants available | Hi,
The smallest Deepseek-v3 you currently offer via the library is deepseek-v3:671b-q4_K_M.
Please can you make smaller quants available, e.g., Q3, to make this model more accessible.
Thanks! | model request | low | Minor |
2,796,984,999 | transformers | Inconsistent output lengths when `max_length=20` is set implicitly vs explicitly in `generate()` | ### System Info
- `transformers` version: 4.49.0.dev0
- Platform: macOS-15.1.1-arm64-arm-64bit
- Python version: 3.11.10
- Huggingface_hub version: 0.27.1
- Safetensors version: 0.5.2
- Accelerate version: 1.2.1
- Accelerate config: not found
- PyTorch version (GPU?): 2.5.1 (False)
- Tensorflow version (GPU?): not installed (NA)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Using distributed or parallel set-up in script?: no
### Who can help?
@gante
@ArthurZucker
Related PR that discusses recent default `max_length`-related changes: #34814.
### Information
- [ ] The official example scripts
- [x] My own modified scripts
### Tasks
- [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)
- [x] My own task or dataset (give details below)
### Reproduction
When using `generate()` with a model that has `generation_config.max_length=20`, the output length differs depending on whether `max_length` is passed explicitly or used implicitly from the `generation_config`.
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Setup from tests/generation/test_utils.py::GenerationIntegrationTests
article = "Today a dragon flew over Paris."
model = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2")
tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2")
input_ids = tokenizer(article, return_tensors="pt").input_ids
# Case 1: Implicit max_length from generation_config
out_gen_implicit = model.generate(input_ids=input_ids)
print(out_gen_implicit.shape[-1]) # 36
# Case 2: Explicit max_length
out_gen_explicit = model.generate(
input_ids=input_ids,
max_length=model.generation_config.max_length
)
print(out_gen_explicit.shape[-1]) # 20
```
In the first case, the generated text is longer than in the second case (36 vs. 20 tokens).
#### Reason and scope
In the first case, `max_length` is overwritten as follows in file `src/transformers/generation/utils.py`, function `_prepare_generated_length`:
```python
if generation_config.max_length == GenerationConfig().max_length:
generation_config.max_length = generation_config.max_length + input_ids_length
```
Since `GenerationConfig().max_length` defaults to 20, the bug only affects models with `generation_config.max_length` set to 20.
### Expected behavior
The calls `model.generate(input_ids=input_ids)` and `model.generate(input_ids=input_ids, max_length=model.generation_config.max_length)` should generate texts of the same length when `generation_config.max_length` is set to 20.
| bug | low | Critical |
2,796,992,863 | godot | [4.4] Camera3D preview doesn't show anything until interact with editor. | ### Tested versions
v4.4.beta1.official [d33da79d3]
### System information
Godot v4.4.beta1 - Android - Single-window, 1 monitor - OpenGL ES 3 (Compatibility) - Adreno (TM) 610 - (8 threads)
### Issue description
After add Camera3D inside a scene camera preview doesn't show anything. It show nothing until I interact with the editor. Also this happen when select the Camera3D node every time.
Video:
https://github.com/user-attachments/assets/37caac9f-d2f8-44b8-8086-d9fb79f80c4a
### Steps to reproduce
- add camera3D inside your tree
- check the preview
### Minimal reproduction project (MRP)
n/a | bug,topic:editor,topic:3d | low | Minor |
2,797,019,519 | flutter | Grade Plugin Flutter | ### Steps to reproduce
flutter build appbundle
FAILURE: Build failed with an exception.
* What went wrong:
A problem occurred configuring root project 'android'.
> Could not resolve all files for configuration ':classpath'.
> Could not find io.flutter:flutter-gradle-plugin:3.0.0.
Searched in the following locations:
- https://dl.google.com/dl/android/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
- https://repo.maven.apache.org/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
- https://storage.googleapis.com/download.flutter.io/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
Required by:
project :
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
> Get more help at https://help.gradle.org.
BUILD FAILED in 4s
Running Gradle task 'bundleRelease'... 5,2s
Gradle task bundleRelease failed with exit code 1
### Actual results
I am trying to build the appbundle file, but I get these errors
### Logs
C:\Users\GGS\vocabulary>flutter pub outdated
Showing outdated packages.
[*] indicates versions that are not the latest available.
Package Name Current Upgradable Resolvable Latest
direct dependencies: all up-to-date.
dev_dependencies: all up-to-date.
transitive dev_dependencies:
boolean_selector *2.1.1 *2.1.1 *2.1.1 2.1.2
fake_async *1.3.1 *1.3.1 *1.3.1 1.3.2
leak_tracker *10.0.7 *10.0.7 *10.0.7 10.0.8
leak_tracker_flutter_testing *3.0.8 *3.0.8 *3.0.8 3.0.9
matcher *0.12.16+1 *0.12.16+1 *0.12.16+1 0.12.17
stack_trace *1.12.0 *1.12.0 *1.12.0 1.12.1
stream_channel *2.1.2 *2.1.2 *2.1.2 2.1.4
test_api *0.7.3 *0.7.3 *0.7.3 0.7.4
vm_service *14.3.0 *14.3.0 *14.3.0 15.0.0
all dependencies are up-to-date.
C:\Users\GGS\vocabulary>flutter clean
C:\Users\GGS\vocabulary>flutter pub get
Resolving dependencies...
Downloading packages...
! async 2.12.0 (overridden)
boolean_selector 2.1.1 (2.1.2 available)
! characters 1.4.0 (overridden)
! clock 1.1.2 (overridden)
! collection 1.19.1 (overridden)
! csslib 1.0.2 (overridden)
fake_async 1.3.1 (1.3.2 available)
! html 0.15.5 (overridden)
leak_tracker 10.0.7 (10.0.8 available)
leak_tracker_flutter_testing 3.0.8 (3.0.9 available)
matcher 0.12.16+1 (0.12.17 available)
! material_color_utilities 0.12.0 (overridden)
! meta 1.16.0 (overridden)
! path 1.9.1 (overridden)
! petitparser 6.1.0 (overridden)
! source_span 1.10.1 (overridden)
stack_trace 1.12.0 (1.12.1 available)
stream_channel 2.1.2 (2.1.4 available)
! string_scanner 1.4.1 (overridden)
! term_glyph 1.2.2 (overridden)
test_api 0.7.3 (0.7.4 available)
vm_service 14.3.0 (15.0.0 available)
Got dependencies!
9 packages have newer versions incompatible with dependency constraints.
Try `flutter pub outdated` for more information.
C:\Users\GGS\vocabulary>flutter build appbundle
FAILURE: Build failed with an exception.
* What went wrong:
A problem occurred configuring root project 'android'.
> Could not resolve all files for configuration ':classpath'.
> Could not find io.flutter:flutter-gradle-plugin:3.0.0.
Searched in the following locations:
- https://dl.google.com/dl/android/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
- https://repo.maven.apache.org/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
- https://storage.googleapis.com/download.flutter.io/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom
Required by:
project :
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
> Get more help at https://help.gradle.org.
BUILD FAILED in 4s
Running Gradle task 'bundleRelease'... 5,2s
Gradle task bundleRelease failed with exit code 1
### Flutter Doctor output
flutter doctor -v
[√] Flutter (Channel stable, 3.27.2, on Microsoft Windows [Versione 10.0.19045.5371], locale it-IT)
• Flutter version 3.27.2 on channel stable at C:\flutter
• Upstream repository https://github.com/flutter/flutter.git
• Framework revision 68415ad1d9 (5 days ago), 2025-01-13 10:22:03 -0800
• Engine revision e672b006cb
• Dart version 3.6.1
• DevTools version 2.40.2
[√] Windows Version (Installed version of Windows is version 10 or higher)
[√] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
• Android SDK at C:\Users\GGS\AppData\Local\Android\Sdk
• Platform android-34, build-tools 34.0.0
• Java binary at: C:\Program Files\Android\Android Studio\jbr\bin\java
• Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-10027231)
• All Android licenses accepted.
[√] Chrome - develop for the web
• Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe
[X] Visual Studio - develop Windows apps
X Visual Studio not installed; this is necessary to develop Windows apps.
Download at https://visualstudio.microsoft.com/downloads/.
Please install the "Desktop development with C++" workload, including all of its default components
[√] Android Studio (version 2022.3)
• Android Studio at C:\Program Files\Android\Android Studio
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-10027231)
[√] Connected device (3 available)
• Windows (desktop) • windows • windows-x64 • Microsoft Windows [Versione 10.0.19045.5371]
• Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83
• Edge (web) • edge • web-javascript • Microsoft Edge 131.0.2903.146
[√] Network resources
• All expected network resources are available.
! Doctor found issues in 1 category. | waiting for customer response,in triage | low | Critical |
2,797,020,257 | godot | GPUParticles3D nodes skip update for one frame whenever a new GPUParticles3D is created | ### Tested versions
Reproducible in: v4.4.beta1.mono.official [d33da79d3]
### System information
Godot v4.4.beta1.mono - Windows 10 (build 19045) - Single-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2060 (NVIDIA; 31.0.15.5275) - Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz (12 threads)
### Issue description
Whenever a scene with a GPUParticles3D node is instantiated, all existing GPUParticles3D nodes skip updates for one frame. Adding the instantiated scene to the scene tree via e.g. `add_sibling` is not necessary for the issue to occur.
The issue is difficult to see with high FPS and/or slow moving particles, but easy to notice with a low FPS cap (e.g. 20) and fast moving particles.
https://github.com/user-attachments/assets/56202ffb-1775-4385-9c0a-3f830713a3ae
The GIF below illustrates the moment of creating a new particle system in slowdown. It can be seen that the frame following the one with the creation of the new GPUParticles3D has the exact same configuration of all particles as the frame prior, but slightly offset due to independent camera movement.

### Steps to reproduce
- In the project settings, set the max_fps option to a low value, e.g. 20. This will make the issue easier to see.
- Create a scene with a GPUParticles3D as a root node. The content is unimportant as long as the particles are visible and not one-shot.
- Create a preview scene.
- Put a Camera3D node in the root of the preview scene, e.g. unrotated, at z=6m. This is for preview purposes.
- Put a Node3D node in the root of the preview scene, e.g. unrotated, at x=-7, y=5. This is a cursor for spawning the particle scene. Attach a script to the node, for example:
```gdscript
extends Node3D
@export var particle_scene: PackedScene
@onready var spawned_particles: Array = []
func _physics_process(_delta):
if Input.is_action_just_pressed("ui_accept"):
if global_position.x < 7.0:
# create a new particle system
var particle = particle_scene.instantiate()
particle.position = global_position
add_sibling(particle)
spawned_particles.push_back(particle)
# push the position
global_position.x += 1.0
else:
# if the position was pushed far enough, reset
for particle in spawned_particles:
particle.queue_free()
spawned_particles.clear()
global_position.x = -7.0
```
- Make sure the particle scene is linked the the cursor node's export variable.
- Run the game in play mode on the preview scene. Pressing Enter should result in the cursor spawning a particle scene and moving.
With the low FPS value, it should be easy to see that every time a particle scene is spawned, all the GPUParticles3D nodes stop updating for exactly one frame. Commenting out the `add_sibling` line shows that the issue still occurs when the new GPUParticles3D node is not added to the scene tree.
If there are other moving nodes on the scene, it can be seen that this is a local issue rather than a global hang-up. Minimal reproduction project comes with slight camera movement to illustrate that the movement of the rest of the game continues smoothly.
### Minimal reproduction project (MRP)
[particletest2_2025-01-18_15-33-43.zip](https://github.com/user-attachments/files/18464787/particletest2_2025-01-18_15-33-43.zip) | bug,confirmed,topic:particles | low | Major |
2,797,025,983 | vscode | properly not working |
Type: <b>Bug</b>
work space is not working in this visual studio code
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Windows_NT x64 6.2.9200
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i5-7300U CPU @ 2.60GHz (4 x 2712)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|undefined|
|Memory (System)|7.88GB (1.42GB free)|
|Process Argv|--crash-reporter-id 678aba3c-fe94-4076-af47-bbf5e066d80f|
|Screen Reader|no|
|VM|0%|
</details><details><summary>Extensions (29)</summary>
Extension|Author (truncated)|Version
---|---|---
python-debug-address-prompt|ale|1.0.1
prettier-vscode|esb|11.0.0
c-cpp-runner|fra|9.4.9
codespaces|Git|1.17.3
cpptask|kay|0.0.1
csdevkit|ms-|1.15.34
csharp|ms-|2.61.28
vscode-dotnet-runtime|ms-|2.2.5
debugpy|ms-|2024.14.0
python|ms-|2024.22.2
vscode-pylance|ms-|2024.12.1
jupyter|ms-|2024.11.0
jupyter-keymap|ms-|1.1.2
jupyter-renderers|ms-|1.0.21
vscode-jupyter-cell-tags|ms-|0.1.9
vscode-jupyter-slideshow|ms-|0.1.6
cmake-tools|ms-|1.19.52
cpptools|ms-|1.22.11
cpptools-extension-pack|ms-|1.3.0
powershell|ms-|2024.4.0
remote-explorer|ms-|0.4.3
remote-server|ms-|1.5.2
java|red|1.38.0
LiveServer|rit|5.7.9
cmake|twx|0.0.17
vscode-lldb|vad|1.11.1
vscode-gradle|vsc|3.16.2024111106
vscode-java-debug|vsc|0.58.1
vscode-java-test|vsc|0.43.0
(1 theme extensions excluded)
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vspor879:30202332
vspor708:30202333
vspor363:30204092
vscod805:30301674
binariesv615:30325510
vsaa593cf:30376535
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
2i9eh265:30646982
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
h48ei257:31000450
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupytercf:31046870
newcmakeconfigv2:31071590
nativerepl1:31139838
pythonrstrctxt:31112756
nativeloc2:31192216
cf971741:31144450
iacca1:31171482
notype1cf:31157160
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
cf1a2727:31215809
6074i472:31201624
dwoutputs:31217127
```
</details>
<!-- generated by issue reporter --> | info-needed,triage-needed | low | Critical |
2,797,036,148 | rust | Missing re-export of `FromCoroutine` in `core::iter` for `from_coroutine` | The `FromCoroutine` type returned by the unstable function `core::iter::from_coroutine` was not publicly re-exported in the `core::iter` module. | C-bug,T-libs | low | Minor |
2,797,037,301 | svelte | Cannot set properties of null (setting 'currentTime'), when transition | ### Describe the bug
When playing audio HTML element with binded currentTime value becomes hidden with transition, the Svelte still tries to set the value of the now missing property (while animated), which results in an error Cannot set properties of null.
### Reproduction
Full example: [REPL](https://svelte.dev/playground/6b0bf9775fe64b0097f0fe18d5eee5ad?version=5.19.0)
Press play and then hide the player
### Logs
```shell
```
### System Info
```shell
Chrome 131
```
### Severity
annoyance | transition/animation | low | Critical |
2,797,039,519 | transformers | Defining LLM Dataset types in Trainers or during Training Workflow | ### Feature request
Hi,
Just wondering if its possible to define a way to define the dataset input labels mapping argument in the transformers.Trainer or something similar?
I understand that the input data labels are dependent on the type of training loss function used and problem for example, the type of defining attention mask ids for Mask learner vs Causal learners etc. Would also be extremely helpful if I can have suggestions on how I can check what dataset column labels the trainer/model will read? -- besides the `help(model.forward)` as this really just shows the model's inputs as it is into the learner.
Any suggestions would be greatly appreciated! thanks!
### Motivation
This problem is giving me clinical depression
### Your contribution
If there's any suggestions in solving this case, I would love to contribute | Feature request | low | Minor |
2,797,048,027 | langchain | OpenAI Assistant V2 Runnable parameters, run_metadata does not exist | ### Checked other resources
- [x] I added a very descriptive title to this issue.
- [x] I searched the LangChain documentation with the integrated search.
- [x] I used the GitHub search to find a similar question and didn't find it.
- [x] I am sure that this is a bug in LangChain rather than my code.
- [x] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package).
### Example Code
```python
from langchain_community.agents.openai_assistant import OpenAIAssistantV2Runnable
assistant = OpenAIAssistantV2Runnable.create_assistant(
name="bug_reproducer",
instructions='Write "A"',
check_every_ms=5_000.0,
tools={},
model="gpt-4o-mini",
)
message = {
"content": "A",
"max_completion_tokens": 10,
"run_metadata": {"key": "value"},
}
resp = assistant.invoke(message)
print(resp)
```
### Error Message and Stack Trace (if applicable)
Traceback (most recent call last):
File "[Redacted]/Langchain Tests/reproducer.py", line 17, in <module>
resp = assistant.invoke(message)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "[Redacted]/lib/python3.12/site-packages/langchain_community/agents/openai_assistant/base.py", line 383, in invoke
raise e
File "[Redacted]/langchain_community/agents/openai_assistant/base.py", line 365, in invoke
run = self._create_thread_and_run(input, thread)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "[Redacted]/langchain_community/agents/openai_assistant/base.py", line 579, in _create_thread_and_run
run = self.client.beta.threads.create_and_run(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "[Redacted]/openai/_utils/_utils.py", line 279, in wrapper
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
TypeError: Threads.create_and_run() got an unexpected keyword argument 'run_metadata'
### Description
Currently the parameters passed down to OpenAI from Langchain are limited to a set of parameters (lately updated in this [PR](https://github.com/langchain-ai/langchain/pull/29207) ).
While testing the usage of the parameters I noticed that one of the parameters already present prior to the PR that could be passed down to openAI is "run_metadata", but verifying on OpenAI library I saw no such parameters, a parameter "metadata" [exists](https://github.com/openai/openai-python/blob/main/src/openai/resources/beta/threads/threads.py#L275) but not the "run_metadata" one.
Going back to [OpenAI 1.1.0](https://github.com/openai/openai-python/blob/v1.1.0/src/openai/resources/beta/threads/threads.py#L201) which introduced assistants I always see "metadata" and not "run_metadata", so that parameter if passed never actually worked).
This could be easily solved by changing the occurences of the parameter "run_metadata" with "metadata" in the allow sets.
But given the current status of the Assistant class and the passage of parameters to OpenAI threads and runs underneath, I was thinking that the current approach with a "set of allowed values" is quite strict, can be edited only via a PR and could break as soon as a change is done on OpenAI python package.
To avoid this kind of issues and keep aligned parameters from OpenAI side and Langchain side the parameters should be passed in a separate object (not in additional keys of the input where we also provide the chat messages) and could be useful to somehow use for example for `thread_create_and_run` method to use the "TypedDict" [`ThreadCreateAndRunParamsBase`](https://github.com/openai/openai-python/blob/main/src/openai/types/beta/thread_create_and_run_params.py#L38) present in OpenAI library.
### System Info
python 3.12
pip freeze
aiohappyeyeballs==2.4.4
aiohttp==3.11.11
aiosignal==1.3.2
annotated-types==0.7.0
anyio==4.8.0
attrs==24.3.0
certifi==2024.12.14
charset-normalizer==3.4.1
dataclasses-json==0.6.7
distro==1.9.0
frozenlist==1.5.0
greenlet==3.1.1
h11==0.14.0
httpcore==1.0.7
httpx==0.28.1
httpx-sse==0.4.0
idna==3.10
jiter==0.8.2
jsonpatch==1.33
jsonpointer==3.0.0
langchain==0.3.14
langchain-community==0.3.14
langchain-core==0.3.30
langchain-text-splitters==0.3.5
langsmith==0.2.11
marshmallow==3.25.1
multidict==6.1.0
mypy-extensions==1.0.0
numpy==2.2.1
openai==1.59.8
orjson==3.10.14
packaging==24.2
propcache==0.2.1
pydantic==2.10.5
pydantic-settings==2.7.1
pydantic_core==2.27.2
python-dotenv==1.0.1
PyYAML==6.0.2
requests==2.32.3
requests-toolbelt==1.0.0
sniffio==1.3.1
SQLAlchemy==2.0.37
tenacity==9.0.0
tqdm==4.67.1
typing-inspect==0.9.0
typing_extensions==4.12.2
urllib3==2.3.0
yarl==1.18.3 | 🤖:bug | low | Critical |
2,797,056,331 | pytorch | Driver Allocated Memory grows unrestricted when using torch.unique on MPS device | ### 🐛 Describe the bug
When using `torch.unique` in a loop on the MPS backend, the memory allocated by the driver grows unrestricted. In my real application that leads to an `RuntimeError: MPS backend out of memory (MPS allocated: 24.00 MB, other allocations: 36.24 GB, max allowed: 36.27 GB)` error late in the training.
I created this minimal example with the same behaviour.
```python
import torch
import gc
def test_operations(iterations: int, shape: tuple[int, int]) -> None:
print(f"PyTorch version: {torch.__version__}")
# Test 1: torch.unique
print("\nTest 1: torch.unique")
x = torch.randint(0, 2, shape, device="mps")
for i in range(iterations):
y = torch.unique(x)
del y
# Empty cache and collect garbage to make sure
torch.mps.empty_cache()
gc.collect()
if i % 10 == 0:
print(
f"Iter {i}: Driver Allocated Memory: {torch.mps.driver_allocated_memory() / (1024**2):.2f}MB, Current Allocated Memory: {torch.mps.current_allocated_memory() / (1024**2):.2f}MB"
)
# Test 2: torch.sort (comparison)
print("\nTest 2: torch.sort")
for i in range(iterations):
y = torch.sort(x)[0]
del y
# Empty cache and collect garbage to make sure
torch.mps.empty_cache()
gc.collect()
if i % 10 == 0:
print(
f"Iter {i}: Driver memory: {torch.mps.driver_allocated_memory() / (1024**2):.2f}MB, Current memory: {torch.mps.current_allocated_memory() / (1024**2):.2f}MB"
)
test_operations(iterations=100, shape=(2000, 10))
```
Results in
```
PyTorch version: 2.5.1
Test 1: torch.unique
Iter 0: Driver Allocated Memory: 18.73MB, Current Allocated Memory: 0.15MB
Iter 10: Driver Allocated Memory: 98.73MB, Current Allocated Memory: 0.15MB
Iter 20: Driver Allocated Memory: 178.73MB, Current Allocated Memory: 0.15MB
Iter 30: Driver Allocated Memory: 258.73MB, Current Allocated Memory: 0.15MB
Iter 40: Driver Allocated Memory: 338.73MB, Current Allocated Memory: 0.15MB
Iter 50: Driver Allocated Memory: 418.73MB, Current Allocated Memory: 0.15MB
Iter 60: Driver Allocated Memory: 578.73MB, Current Allocated Memory: 0.15MB
Iter 70: Driver Allocated Memory: 738.72MB, Current Allocated Memory: 0.15MB
Iter 80: Driver Allocated Memory: 898.72MB, Current Allocated Memory: 0.15MB
Iter 90: Driver Allocated Memory: 1058.72MB, Current Allocated Memory: 0.15MB
Test 2: torch.sort
Iter 0: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 10: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 20: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 30: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 40: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 50: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 60: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 70: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 80: Driver memory: 1202.72MB, Current memory: 0.15MB
Iter 90: Driver memory: 1202.72MB, Current memory: 0.15MB
```
Showing the increase in the driver allocated memory when using `torch.unique` but not when using another function like `torch.sort`. I just used `torch.sort` as comparison here.
Is this behaviour expected?
### Versions
Collecting environment information...
PyTorch version: 2.5.1
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: macOS 15.2 (arm64)
GCC version: Could not collect
Clang version: 16.0.0 (clang-1600.0.26.6)
CMake version: Could not collect
Libc version: N/A
Python version: 3.11.11 (main, Dec 3 2024, 17:20:40) [Clang 16.0.0 (clang-1600.0.26.4)] (64-bit runtime)
Python platform: macOS-15.2-arm64-arm-64bit
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Apple M2 Pro
Versions of relevant libraries:
[pip3] torch==2.5.1
[conda] Could not collect
cc @kulinseth @albanD @malfet @DenisVieriu97 @jhavukainen | module: memory usage,triaged,module: mps | low | Critical |
2,797,056,878 | rust | Tracking issue for release notes of #133925: disallow `repr()` on invalid items |
This issue tracks the release notes text for #133925.
### Steps
- [ ] Proposed text is drafted by PR author (or team) making the noteworthy change.
- [ ] Issue is nominated for release team review of clarity for wider audience.
- [ ] Release team includes text in release notes/blog posts.
### Release notes text
The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing).
````markdown
# Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...)
- [disallow `repr()` on invalid items](https://github.com/rust-lang/rust/pull/133925)
````
> [!TIP]
> Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use.
> The category will be de-duplicated with all the other ones by the release team.
>
> *More than one section can be included if needed.*
### Release blog section
If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section.
*Otherwise leave it empty.*
````markdown
````
cc @folkertdev, @compiler-errors -- origin issue/PR authors and assignees for starting to draft text
| T-lang,relnotes,needs-triage,relnotes-tracking-issue | low | Critical |
2,797,065,765 | rust | Tracking issue for release notes of #133889: Consider fields to be inhabited if they are unstable |
This issue tracks the release notes text for #133889.
### Steps
- [ ] Proposed text is drafted by PR author (or team) making the noteworthy change.
- [ ] Issue is nominated for release team review of clarity for wider audience.
- [ ] Release team includes text in release notes/blog posts.
### Release notes text
The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing).
````markdown
# Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...)
- [Consider fields to be inhabited if they are unstable](https://github.com/rust-lang/rust/pull/133889)
````
> [!TIP]
> Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use.
> The category will be de-duplicated with all the other ones by the release team.
>
> *More than one section can be included if needed.*
### Release blog section
If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section.
*Otherwise leave it empty.*
````markdown
````
cc @compiler-errors, @Nadrieril -- origin issue/PR authors and assignees for starting to draft text
| T-lang,relnotes,needs-triage,relnotes-tracking-issue | low | Critical |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.