id
int64
393k
2.82B
repo
stringclasses
68 values
title
stringlengths
1
936
body
stringlengths
0
256k
labels
stringlengths
2
508
priority
stringclasses
3 values
severity
stringclasses
3 values
2,796,110,878
go
net: TestUDPServer/2 failures
``` #!watchflakes default <- pkg == "net" && test == "TestUDPServer/2" ``` Issue created automatically to collect these failures. Example ([log](https://ci.chromium.org/b/8725537367198926609)): === RUN TestUDPServer/2 server_test.go:258: udp [::ffff:0.0.0.0]:0<-127.0.0.1 server_test.go:322: server: read udp [::]:56739: i/o timeout server_test.go:315: client: read udp 127.0.0.1:50097: i/o timeout --- FAIL: TestUDPServer/2 (3600.00s) — [watchflakes](https://go.dev/wiki/Watchflakes)
NeedsInvestigation
low
Critical
2,796,113,577
next.js
Prefetched PPR page segments stay suspended when navigated to after a rebuild
### Link to the code that reproduces this issue https://github.com/bradleydsmith/reproduction-next-prefetch-rebuild-ppr-suspended ### To Reproduce 1. Build the application with npm run build 2. Start the server with npm run start 3. Navigate to http://localhost:3000 and do not click anything or close the page 4. Stop the server 5. Run npm run build 6. Start the server again with npm run start 7. Click the Prefetched link on the page ### Current vs. Expected behavior When following the steps the suspended segment stays suspended and does not get resolved and rendered. I expected the page to do an MPA navigation and reload due to detecting a change in the response build ID ### Provide environment information ```bash Operating System: Platform: darwin Arch: x64 Version: Darwin Kernel Version 23.5.0: Wed May 1 20:09:52 PDT 2024; root:xnu-10063.121.3~5/RELEASE_X86_64 Available memory (MB): 8192 Available CPU cores: 8 Binaries: Node: 20.9.0 npm: 10.1.0 Yarn: N/A pnpm: 9.6.0 Relevant Packages: next: 15.2.0-canary.14 // Latest available version is detected (15.2.0-canary.14). eslint-config-next: N/A react: 19.0.0 react-dom: 19.0.0 typescript: 5.7.3 Next.js Config: output: N/A ``` ### Which area(s) are affected? (Select all that apply) dynamicIO, Partial Prerendering (PPR) ### Which stage(s) are affected? (Select all that apply) Vercel (Deployed), next start (local), next build (local) ### Additional context The issue is being caused by an incorrect check in listenForDynamicRequest. The issue occurs after a rebuild of the server, this could be a local rebuild and restart of the server but also affects a deployment to Vercel without skew protection. Also note how the issue does not happen on the Not Prefetched link, it only affects prefetched pages
Partial Prerendering (PPR),dynamicIO
low
Minor
2,796,115,021
rust
new build_helper::compiletest module for safer invocations of compiletest by bootstrap
currently bootstrap just uses strings to represent various compiletest arguments, such as mode. obviously it would be better to use enums for these, as they are much less vulnerable to typos. `compiletest` already has enum definitions for these (in `common.rs`), and the two already have a shared dependency (`build_helper`), and `build_helper` already has access to `serde`. i would be willing to work on this refactor, if people think it would be a good idea.
C-cleanup,T-bootstrap,A-compiletest
low
Minor
2,796,117,237
rust
ICE: polonius: `None` in region infer with unconditional panic
<!-- ICE: Rustc ./a.rs '--edition=2024 -Zpolonius=next -ooutputfile -Zdump-mir-dir=dir' 'thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10: 'called `Option::unwrap()` on a `None` value'', 'thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10: 'called `Option::unwrap()` on a `None` value'' File: /tmp/im/a.rs --> not sure why this only panics in edition 2024 🤔 auto-reduced (treereduce-rust): ````rust //@compile-flags: --edition=2024 -Zpolonius=next fn main() { &{ [1, 2, 3][4] }; } ```` original: ````rust //@ build-pass //@ ignore-pass (test emits codegen-time warnings and verifies that they are not errors) #![warn(unconditional_panic)] fn main() { &{ [1, 2, 3][4] }; //~^ WARN operation will panic } ```` Version information ```` rustc 1.86.0-nightly (bcd0683e5 2025-01-17) binary: rustc commit-hash: bcd0683e5dce1945b5d940714742e7502883bb5c commit-date: 2025-01-17 host: x86_64-unknown-linux-gnu release: 1.86.0-nightly LLVM version: 19.1.7 ```` Possibly related line of code: https://github.com/rust-lang/rust/blob/bcd0683e5dce1945b5d940714742e7502883bb5c/compiler/rustc_borrowck/src/region_infer/mod.rs#L1882-L1894 Command: `/home/matthias/.rustup/toolchains/master/bin/rustc --edition=2024 -Zpolonius=next` <details><summary><strong>Program output</strong></summary> <p> ``` thread 'rustc' panicked at compiler/rustc_borrowck/src/region_infer/mod.rs:1888:10: called `Option::unwrap()` on a `None` value stack backtrace: 0: 0x7077524fb0ca - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::h245d3d13c9e117ef 1: 0x707752c12d26 - core::fmt::write::ha5c79450f6a60aba 2: 0x707753b95891 - std::io::Write::write_fmt::hc0bb623be5e87022 3: 0x7077524faf22 - std::sys::backtrace::BacktraceLock::print::h5f0375c539e6bbfe 4: 0x7077524fd3a2 - std::panicking::default_hook::{{closure}}::hfaae0dc7005f1f46 5: 0x7077524fd22a - std::panicking::default_hook::had3887552141764d 6: 0x70775165e33b - std[970739192a5a8910]::panicking::update_hook::<alloc[c190f769f0908da9]::boxed::Box<rustc_driver_impl[98a1c83b91f31d14]::install_ice_hook::{closure#1}>>::{closure#0} 7: 0x7077524fdee3 - std::panicking::rust_panic_with_hook::h0f6e30840497503a 8: 0x7077524fdba6 - std::panicking::begin_panic_handler::{{closure}}::h00d1f33c4abe282b 9: 0x7077524fb5a9 - std::sys::backtrace::__rust_end_short_backtrace::h7f80f131d37aefa4 10: 0x7077524fd89d - rust_begin_unwind 11: 0x70774f1a8530 - core::panicking::panic_fmt::h313dbd2ed26f8657 12: 0x70774fcc082c - core::panicking::panic::h6bed83e7eae20130 13: 0x70775067d5a9 - core::option::unwrap_failed::h5e1ae1d3b98bd9cb 14: 0x70775143be47 - <rustc_borrowck[4bd42f5d1352065e]::region_infer::RegionInferenceContext>::find_sub_region_live_at 15: 0x707750fd0fb4 - <rustc_borrowck[4bd42f5d1352065e]::MirBorrowckCtxt>::explain_why_borrow_contains_point 16: 0x707750e18b6b - <rustc_borrowck[4bd42f5d1352065e]::MirBorrowckCtxt>::report_borrowed_value_does_not_live_long_enough 17: 0x707753cf134f - rustc_borrowck[4bd42f5d1352065e]::do_mir_borrowck 18: 0x707753cc352f - rustc_query_impl[a1940b4b03f25947]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[a1940b4b03f25947]::query_impl::mir_borrowck::dynamic_query::{closure#2}::{closure#0}, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 8usize]>> 19: 0x707752f2260e - rustc_query_system[f87afe778d5ed02f]::query::plumbing::try_execute_query::<rustc_query_impl[a1940b4b03f25947]::DynamicConfig<rustc_data_structures[7c7a3305c98b4df4]::vec_cache::VecCache<rustc_span[e90766ed4fb19497]::def_id::LocalDefId, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 8usize]>, rustc_query_system[f87afe778d5ed02f]::dep_graph::graph::DepNodeIndex>, false, false, false>, rustc_query_impl[a1940b4b03f25947]::plumbing::QueryCtxt, false> 20: 0x707752f22113 - rustc_query_impl[a1940b4b03f25947]::query_impl::mir_borrowck::get_query_non_incr::__rust_end_short_backtrace 21: 0x707752f16fb0 - rustc_interface[f2cc204a1e18dc42]::passes::run_required_analyses 22: 0x707753b9109e - rustc_interface[f2cc204a1e18dc42]::passes::analysis 23: 0x707753b9106f - rustc_query_impl[a1940b4b03f25947]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[a1940b4b03f25947]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 0usize]>> 24: 0x707753be8355 - rustc_query_system[f87afe778d5ed02f]::query::plumbing::try_execute_query::<rustc_query_impl[a1940b4b03f25947]::DynamicConfig<rustc_query_system[f87afe778d5ed02f]::query::caches::SingleCache<rustc_middle[1a423a87a2792eea]::query::erase::Erased<[u8; 0usize]>>, false, false, false>, rustc_query_impl[a1940b4b03f25947]::plumbing::QueryCtxt, false> 25: 0x707753be808e - rustc_query_impl[a1940b4b03f25947]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace 26: 0x707753c5c2e9 - rustc_interface[f2cc204a1e18dc42]::passes::create_and_enter_global_ctxt::<core[31d5326fba56f7df]::option::Option<rustc_interface[f2cc204a1e18dc42]::queries::Linker>, rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}::{closure#2}>::{closure#2}::{closure#0} 27: 0x707753c4f4d6 - rustc_interface[f2cc204a1e18dc42]::interface::run_compiler::<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1} 28: 0x707753a92c07 - std[970739192a5a8910]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_with_globals<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_pool_with_globals<rustc_interface[f2cc204a1e18dc42]::interface::run_compiler<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()> 29: 0x707753a928d9 - <<std[970739192a5a8910]::thread::Builder>::spawn_unchecked_<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_with_globals<rustc_interface[f2cc204a1e18dc42]::util::run_in_thread_pool_with_globals<rustc_interface[f2cc204a1e18dc42]::interface::run_compiler<(), rustc_driver_impl[98a1c83b91f31d14]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>::{closure#1} as core[31d5326fba56f7df]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0} 30: 0x707753a9206b - std::sys::pal::unix::thread::Thread::new::thread_start::h84cef28dceb72f89 31: 0x70774dea339d - <unknown> 32: 0x70774df2849c - <unknown> 33: 0x0 - <unknown> error: the compiler unexpectedly panicked. this is a bug. note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md note: please make sure that you have updated to the latest nightly note: rustc 1.86.0-nightly (bcd0683e5 2025-01-17) running on x86_64-unknown-linux-gnu note: compiler flags: -Z polonius=next -Z dump-mir-dir=dir query stack during panic: #0 [mir_borrowck] borrow-checking `main` #1 [analysis] running analysis passes on this crate end of query stack ``` </p> </details> <!-- query stack: #0 [mir_borrowck] borrow-checking `main` #1 [analysis] running analysis passes on this crate -->
A-diagnostics,I-ICE,C-bug,NLL-polonius,T-types
low
Critical
2,796,123,769
rust
universally directly depend on serde_derive instead of accessing it through serde.
bootstrap does this in order to flatten its dependencies and improve build times, but since other crates (like `compiletest` and `rustdoc`) use the `derive` feature flag, i believe this means `serde` has to be built twice to use any of these programs. wouldn't it be best to always use the same set of feature flags to serde?
C-cleanup,T-bootstrap
low
Minor
2,796,174,444
pytorch
`torch.onnx.export` (dynamo=False) fails with uninformative error when exporting `apply_rotary_pos_emb`/`repeat_interleave`
### 🐛 Describe the bug When attempting to export a HF transformers model that performs `apply_rotary_pos_emb` (definition below), an uninformative error is thrown: <details> <summary>See reproduction code</summary> ```py import torch def rotate_half(x): """Rotates half the hidden dims of the input.""" x1 = x[..., 0::2] x2 = x[..., 1::2] return torch.stack((-x2, x1), dim=-1).flatten(-2) def apply_rotary_pos_emb(q, k, cos, sin, position_ids=None, unsqueeze_dim=1): """Applies Rotary Position Embedding to the query and key tensors. Args: q (`torch.Tensor`): The query tensor. k (`torch.Tensor`): The key tensor. cos (`torch.Tensor`): The cosine part of the rotary embedding. sin (`torch.Tensor`): The sine part of the rotary embedding. position_ids (`torch.Tensor`, *optional*): Deprecated and unused. unsqueeze_dim (`int`, *optional*, defaults to 1): The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2. Returns: `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding. """ cos = cos.unsqueeze(unsqueeze_dim) sin = sin.unsqueeze(unsqueeze_dim) # Interleave them instead of usual shape cos = cos[..., : cos.shape[-1] // 2].repeat_interleave(2, dim=-1) sin = sin[..., : sin.shape[-1] // 2].repeat_interleave(2, dim=-1) # Keep half or full tensor for later concatenation rotary_dim = cos.shape[-1] q_rot, q_pass = q[..., :rotary_dim], q[..., rotary_dim:] k_rot, k_pass = k[..., :rotary_dim], k[..., rotary_dim:] # Apply rotary embeddings on the first half or full tensor q_embed = (q_rot * cos) + (rotate_half(q_rot) * sin) k_embed = (k_rot * cos) + (rotate_half(k_rot) * sin) # Concatenate back to full shape q_embed = torch.cat([q_embed, q_pass], dim=-1) k_embed = torch.cat([k_embed, k_pass], dim=-1) return q_embed, k_embed class Model(torch.nn.Module): def __init__(self): super().__init__() def forward(self, q, k, cos, sin): return apply_rotary_pos_emb(q, k, cos, sin) model = Model() torch.manual_seed(0) q = torch.randn(2, 2, 40, 32) k = torch.randn(2, 2, 40, 32) cos = torch.randn(1, 40, 28) sin = torch.randn(1, 40, 28) output = model(q, k, cos, sin) assert output[0].shape == q.shape assert output[1].shape == k.shape assert abs(output[0].mean().item() - -0.022467557340860367) < 1e-8 assert abs(output[1].mean().item() - 0.0071802930906414986) < 1e-8 torch.onnx.export( model, (q, k, cos, sin), "model.onnx", input_names=["q", "k", "cos", "sin"], output_names=["q_embed", "k_embed"], opset_version=18, dynamo=False, # True works, False is broken ) ``` </details> produces: ``` Traceback (most recent call last): File "/workspaces/optimum/o/rot.py", line 74, in <module> torch.onnx.export( File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/__init__.py", line 375, in export export( File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 502, in export _export( File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1564, in _export graph, params_dict, torch_out = _model_to_graph( ^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1117, in _model_to_graph graph = _optimize_graph( ^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 639, in _optimize_graph graph = _C._jit_pass_onnx(graph, operator_export_type) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/utils.py", line 1836, in _run_symbolic_function return symbolic_fn(graph_context, *inputs, **attrs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_helper.py", line 369, in wrapper return fn(g, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_opset11.py", line 519, in cat return opset9.cat(g, tensor_list, dim) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_helper.py", line 281, in wrapper return fn(g, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/python/3.12.1/lib/python3.12/site-packages/torch/onnx/symbolic_opset9.py", line 563, in cat assert all( AssertionError ``` The issue seems to be caused by the `repeat_interleave` call. However, the error message does not describe this. NOTE: The model exports correctly with dynamo=True, and I understand that the previous exporter is not being worked on, so feel free to close this issue if needed. I raise it for these reasons: - [Optimum](http://github.com/huggingface/optimum) still uses the previous exporter, and so this issue is present. - I haven't been able to find this issue reported, so I'm adding it so I can link to it in my workaround (rewriting the `repeat_interleave` function). - To possibly get a more informative error (since the current one doesn't indicate what the problem is) ### Versions Collecting environment information... PyTorch version: 2.5.1+cu124 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.6 LTS (x86_64) GCC version: (Ubuntu 9.4.0-1ubuntu1~20.04.2) 9.4.0 Clang version: 10.0.0-4ubuntu1 CMake version: version 3.16.3 Libc version: glibc-2.31 Python version: 3.12.1 (main, Sep 30 2024, 17:05:21) [GCC 9.4.0] (64-bit runtime) Python platform: Linux-6.5.0-1025-azure-x86_64-with-glibc2.31 Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian Address sizes: 48 bits physical, 48 bits virtual CPU(s): 2 On-line CPU(s) list: 0,1 Thread(s) per core: 2 Core(s) per socket: 1 Socket(s): 1 NUMA node(s): 1 Vendor ID: AuthenticAMD CPU family: 25 Model: 1 Model name: AMD EPYC 7763 64-Core Processor Stepping: 1 CPU MHz: 3243.502 BogoMIPS: 4890.85 Virtualization: AMD-V Hypervisor vendor: Microsoft Virtualization type: full L1d cache: 32 KiB L1i cache: 32 KiB L2 cache: 512 KiB L3 cache: 32 MiB NUMA node0 CPU(s): 0,1 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode Vulnerability Spec store bypass: Vulnerable Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Retpolines; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl tsc_reliable nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext invpcid_single vmmcall fsgsbase bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr rdpru arat npt nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold v_vmsave_vmload umip vaes vpclmulqdq rdpid fsrm Versions of relevant libraries: [pip3] numpy==2.0.2 [pip3] nvidia-cublas-cu12==12.4.5.8 [pip3] nvidia-cuda-cupti-cu12==12.4.127 [pip3] nvidia-cuda-nvrtc-cu12==12.4.127 [pip3] nvidia-cuda-runtime-cu12==12.4.127 [pip3] nvidia-cudnn-cu12==9.1.0.70 [pip3] nvidia-cufft-cu12==11.2.1.3 [pip3] nvidia-curand-cu12==10.3.5.147 [pip3] nvidia-cusolver-cu12==11.6.1.9 [pip3] nvidia-cusparse-cu12==12.3.1.170 [pip3] nvidia-nccl-cu12==2.21.5 [pip3] nvidia-nvjitlink-cu12==12.4.127 [pip3] nvidia-nvtx-cu12==12.4.127 [pip3] onnx==1.17.0 [pip3] onnx-graphsurgeon==0.5.2 [pip3] onnxconverter-common==1.14.0 [pip3] onnxruntime==1.20.1 [pip3] onnxscript==0.1.0.dev20241219 [pip3] onnxslim==0.1.42 [pip3] torch==2.5.1 [pip3] torchvision==0.20.1 [pip3] triton==3.1.0 [conda] No relevant packages
module: onnx,triaged,OSS contribution wanted
low
Critical
2,796,188,005
rust
ICE: `cannot convert to a region vid`
<!-- ICE: Rustc ./a.rs '-Zincremental-verify-ich=yes -Cincremental=<dir> -Cdebuginfo=2 -Clink-dead-code=true -Zvalidate-mir --edition=2024' 'error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid', 'error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid' File: /tmp/im2/a.rs --> snippet: ````rust struct A<'a> { x: Vec<A> = Vec::new(), } ```` Version information ```` rustc 1.86.0-nightly (bcd0683e5 2025-01-17) binary: rustc commit-hash: bcd0683e5dce1945b5d940714742e7502883bb5c commit-date: 2025-01-17 host: x86_64-unknown-linux-gnu release: 1.86.0-nightly LLVM version: 19.1.7 ```` Possibly related line of code: https://github.com/rust-lang/rust/blob/bcd0683e5dce1945b5d940714742e7502883bb5c/compiler/rustc_borrowck/src/universal_regions.rs#L882-L894 Command: `/home/matthias/.rustup/toolchains/master/bin/rustc` <details><summary><strong>Program output</strong></summary> <p> ``` error[E0106]: missing lifetime specifier --> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:2:12 | 2 | x: Vec<A> = Vec::new(), | ^ expected named lifetime parameter | help: consider using the `'a` lifetime | 2 | x: Vec<A<'a>> = Vec::new(), | ++++ error[E0658]: default values on fields are experimental --> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:2:14 | 2 | x: Vec<A> = Vec::new(), | ^^^^^^^^^^^^^ | = note: see issue #132162 <https://github.com/rust-lang/rust/issues/132162> for more information = help: add `#![feature(default_field_values)]` to the crate attributes to enable = note: this compiler was built on 2025-01-17; consider upgrading it if it is out of date error[E0601]: `main` function not found in crate `mvce` --> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:3:2 | 3 | } | ^ consider adding a `main` function to `/tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs` error[E0392]: lifetime parameter `'a` is never used --> /tmp/icemaker_global_tempdir.65vzQ3rSPMlN/rustc_testrunner_tmpdir_reporting.9LgfFhJdaXkA/mvce.rs:1:10 | 1 | struct A<'a> { | ^^ unused lifetime parameter | = help: consider removing `'a`, referring to it in a field, or using a marker such as `PhantomData` error: internal compiler error: compiler/rustc_borrowck/src/universal_regions.rs:888:36: cannot convert `'a/#0` to a region vid thread 'rustc' panicked at compiler/rustc_borrowck/src/universal_regions.rs:888:36: Box<dyn Any> stack backtrace: 0: 0x730ea98f502a - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::h4fb5b0de48364101 1: 0x730eaa012d66 - core::fmt::write::h6eb54c31028066ca 2: 0x730eaafa43d1 - std::io::Write::write_fmt::hd9e9dfb2c5e220cc 3: 0x730ea98f4e82 - std::sys::backtrace::BacktraceLock::print::ha2ddb07494fe7b3b 4: 0x730ea98f7302 - std::panicking::default_hook::{{closure}}::h4eed2bdbab3e3d9c 5: 0x730ea98f718a - std::panicking::default_hook::h520b49f6397a0732 6: 0x730ea8a58e1b - std[3244471bfd6eaa61]::panicking::update_hook::<alloc[8ba9d48c156baf43]::boxed::Box<rustc_driver_impl[df45097da6562dc6]::install_ice_hook::{closure#1}>>::{closure#0} 7: 0x730ea98f7e43 - std::panicking::rust_panic_with_hook::ha20255168d758ad1 8: 0x730ea8a94761 - std[3244471bfd6eaa61]::panicking::begin_panic::<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug>::{closure#0} 9: 0x730ea8a89666 - std[3244471bfd6eaa61]::sys::backtrace::__rust_end_short_backtrace::<std[3244471bfd6eaa61]::panicking::begin_panic<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug>::{closure#0}, !> 10: 0x730ea8a89423 - std[3244471bfd6eaa61]::panicking::begin_panic::<rustc_errors[af9a6eaf1dbd0244]::ExplicitBug> 11: 0x730ea8a9e691 - <rustc_errors[af9a6eaf1dbd0244]::diagnostic::BugAbort as rustc_errors[af9a6eaf1dbd0244]::diagnostic::EmissionGuarantee>::emit_producing_guarantee 12: 0x730ea907bc53 - rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt::<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0} 13: 0x730ea90615da - rustc_middle[352907c3633efb2c]::ty::context::tls::with_opt::<rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0}, !>::{closure#0} 14: 0x730ea906146b - rustc_middle[352907c3633efb2c]::ty::context::tls::with_context_opt::<rustc_middle[352907c3633efb2c]::ty::context::tls::with_opt<rustc_middle[352907c3633efb2c]::util::bug::opt_span_bug_fmt<rustc_span[2bc6ccfbd90abceb]::span_encoding::Span>::{closure#0}, !>::{closure#0}, !> 15: 0x730ea6f596b0 - rustc_middle[352907c3633efb2c]::util::bug::bug_fmt 16: 0x730eab54beb3 - <rustc_borrowck[55cebfbcecacd79c]::universal_regions::UniversalRegionIndices>::to_region_vid.cold 17: 0x730eaa060745 - <rustc_middle[352907c3633efb2c]::ty::Ty as rustc_type_ir[ba8c709077c11182]::fold::TypeSuperFoldable<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>::try_super_fold_with::<rustc_type_ir[ba8c709077c11182]::fold::RegionFolder<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>> 18: 0x730eaa05f068 - <rustc_middle[352907c3633efb2c]::ty::Ty as rustc_type_ir[ba8c709077c11182]::fold::TypeSuperFoldable<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>>::try_super_fold_with::<rustc_type_ir[ba8c709077c11182]::fold::RegionFolder<rustc_middle[352907c3633efb2c]::ty::context::TyCtxt>> 19: 0x730eab0c21f6 - rustc_borrowck[55cebfbcecacd79c]::do_mir_borrowck 20: 0x730eab0bfbef - rustc_query_impl[43ade5033af9534a]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[43ade5033af9534a]::query_impl::mir_borrowck::dynamic_query::{closure#2}::{closure#0}, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 8usize]>> 21: 0x730eaa33c04e - rustc_query_system[eb70713c06242342]::query::plumbing::try_execute_query::<rustc_query_impl[43ade5033af9534a]::DynamicConfig<rustc_data_structures[dc2c66faf0aa2c5f]::vec_cache::VecCache<rustc_span[2bc6ccfbd90abceb]::def_id::LocalDefId, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 8usize]>, rustc_query_system[eb70713c06242342]::dep_graph::graph::DepNodeIndex>, false, false, false>, rustc_query_impl[43ade5033af9534a]::plumbing::QueryCtxt, false> 22: 0x730eaa33bb53 - rustc_query_impl[43ade5033af9534a]::query_impl::mir_borrowck::get_query_non_incr::__rust_end_short_backtrace 23: 0x730eaa3309f0 - rustc_interface[508537c9f518ed83]::passes::run_required_analyses 24: 0x730eaaf9fade - rustc_interface[508537c9f518ed83]::passes::analysis 25: 0x730eaaf9faaf - rustc_query_impl[43ade5033af9534a]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[43ade5033af9534a]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 0usize]>> 26: 0x730eaaf5a9d5 - rustc_query_system[eb70713c06242342]::query::plumbing::try_execute_query::<rustc_query_impl[43ade5033af9534a]::DynamicConfig<rustc_query_system[eb70713c06242342]::query::caches::SingleCache<rustc_middle[352907c3633efb2c]::query::erase::Erased<[u8; 0usize]>>, false, false, false>, rustc_query_impl[43ade5033af9534a]::plumbing::QueryCtxt, false> 27: 0x730eaaf5a70e - rustc_query_impl[43ade5033af9534a]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace 28: 0x730eab0391e9 - rustc_interface[508537c9f518ed83]::passes::create_and_enter_global_ctxt::<core[24a92a8ec63ee5f4]::option::Option<rustc_interface[508537c9f518ed83]::queries::Linker>, rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}::{closure#2}>::{closure#2}::{closure#0} 29: 0x730eab02c416 - rustc_interface[508537c9f518ed83]::interface::run_compiler::<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1} 30: 0x730eaae73a07 - std[3244471bfd6eaa61]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[508537c9f518ed83]::util::run_in_thread_with_globals<rustc_interface[508537c9f518ed83]::util::run_in_thread_pool_with_globals<rustc_interface[508537c9f518ed83]::interface::run_compiler<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()> 31: 0x730eaae736d9 - <<std[3244471bfd6eaa61]::thread::Builder>::spawn_unchecked_<rustc_interface[508537c9f518ed83]::util::run_in_thread_with_globals<rustc_interface[508537c9f518ed83]::util::run_in_thread_pool_with_globals<rustc_interface[508537c9f518ed83]::interface::run_compiler<(), rustc_driver_impl[df45097da6562dc6]::run_compiler::{closure#0}>::{closure#1}, ()>::{closure#0}, ()>::{closure#0}::{closure#0}, ()>::{closure#1} as core[24a92a8ec63ee5f4]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0} 32: 0x730eaae72e6f - std::sys::pal::unix::thread::Thread::new::thread_start::h5d259864ca2af1e4 33: 0x730ea52a339d - <unknown> 34: 0x730ea532849c - <unknown> 35: 0x0 - <unknown> note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md note: please make sure that you have updated to the latest nightly note: please attach the file at `/tmp/im2/rustc-ice-2025-01-17T19_45_25-2933219.txt` to your bug report query stack during panic: #0 [mir_borrowck] borrow-checking `A::x::{constant#0}` #1 [analysis] running analysis passes on this crate end of query stack error: aborting due to 5 previous errors Some errors have detailed explanations: E0106, E0392, E0601, E0658. For more information about an error, try `rustc --explain E0106`. ``` </p> </details> <!-- query stack: note: compiler flags: -Z incremental-verify-ich=yes -C incremental=[REDACTED] -C debuginfo=2 -C link-dead-code=true -Z validate-mir #0 [mir_borrowck] borrow-checking `A::x::{constant#0}` #1 [analysis] running analysis passes on this crate -->
I-ICE,T-compiler,C-bug
low
Critical
2,796,198,136
deno
Error when using TS Decorators
Version: Deno 2.1.6 Error when using argument decorators ![Image](https://github.com/user-attachments/assets/6bc5e380-6dde-4770-9ace-cbb72178a9f6)
needs investigation
low
Critical
2,796,212,730
rust
Change in leaf function causes unrelated compilation error in root function
## Description I encountered an issue where changing the leaf function causes a compilation error in the root function, but the change doesn't affect types or auto traits of the function. Commenting out `router()` makes code to compile. ## Minimal Example ```rust use axum::routing::{post, Router}; use futures_util::StreamExt; use std::collections::HashMap; use std::sync::Arc; struct Transaction { f1: (), f2: (), } pub(crate) struct SplitAddress(()); async fn handler_inner(ctx: Context) { let transactions = vec![]; // This causes the lifetime issue: let iter = transactions .iter() .map(|x: &Transaction| (x.f1, SplitAddress(x.f2))); // This works: // let iter: Vec<_> = transactions // .iter() // .map(|x: &Transaction| (x.f1, SplitAddress(x.f2))) // .collect(); ctx.0.streaming_work(iter.into_iter()).await; } async fn handler() -> Result<String, String> { let ctx = Context(Arc::new(Service())); handler_inner(ctx).await; Ok("ok".to_string()) } // or comment out router pub async fn router() -> Router { Router::new().route("/bla", post(handler)) } #[derive(Clone)] pub struct Context(Arc<Service>); struct Service(); impl Service { pub(crate) async fn streaming_work( &self, data: impl Iterator<Item = ((), SplitAddress)>, ) -> HashMap<(), Vec<String>> { futures_util::stream::iter(data) .map(|_| async move { todo!() }) .buffer_unordered(100) .filter_map(|x| async move { x }) .collect() .await } } ``` [repo](https://github.com/0xdeafbeef/rustc-issue) ## Error Message ``` error: implementation of `FnOnce` is not general enough --> src/lib.rs:39:33 | 39 | Router::new().route("/bla", post(handler)) | ^^^^^^^^^^^^^ implementation of `FnOnce` is not general enough | = note: closure with signature `fn(&'0 Transaction) -> ((), SplitAddress)` must implement `FnOnce<(&'1 Transaction,)>`, for any two lifetimes `'0` and `'1`... = note: ...but it actually implements `FnOnce<(&Transaction,)>` error: could not compile `rust-issue` (lib) due to 1 previous error ```
A-lifetimes,A-trait-system,needs-triage
low
Critical
2,796,227,000
pytorch
torch._C._IncludeDispatchKeyGuard is very broken?
```py import torch print(torch._C._meta_in_tls_dispatch_include()) with torch._C._IncludeDispatchKeyGuard(torch.DispatchKey.Meta): print(torch._C._meta_in_tls_dispatch_include()) print(torch._C._meta_in_tls_dispatch_include()) ``` prints False, True, True, which is completely bogus
triaged,module: dispatch
low
Critical
2,796,232,607
godot
Camera3D preview may show temporary visual glitch when selected
### Tested versions - Godot v4.4.beta1 - related https://github.com/godotengine/godot/pull/90778 ### System information Windows 11 (build 22631) - Single-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2070 (NVIDIA; 32.0.15.6614) - Intel(R) Core(TM) i7-10750H CPU @ 2.60GHz (12 threads) ### Issue description When there are multiple cameras and you switch between them, there may be a short visual glitch in the camera preview. It doesn't really need multiple cameras to happen, but it happens less often. https://github.com/user-attachments/assets/92811fc9-99c5-4873-92e3-15a639ddc09d Maybe related: the preview updates late in some cases, see https://github.com/godotengine/godot/pull/90778#issuecomment-2395606441 The preview should not show anything until it is ready. ### Steps to reproduce 1. Add 2 Camera3Ds to a scene. 2. Repeatedly switch between them while watching the preview. 3. You may see visual glitches. ### Minimal reproduction project (MRP) Not really needed but here: [camera_preview_multiple_2025-01-17_15-18-50.zip](https://github.com/user-attachments/files/18459220/camera_preview_multiple_2025-01-17_15-18-50.zip)
bug,topic:editor
low
Minor
2,796,236,006
vscode
Remote SSH freezing when loading extensions. Enabling remote.downloadExtensionsLocally did not help
Type: <b>Bug</b> In the server output I'm seeing: 2025-01-17 14:08:27.683 [error] [network] #1: https://main.vscode-cdn.net/extensions/marketplace.json - error GET read ETIMEDOUT Yes I know remote SSH can be a problem if the remote server is behind a firewall. However I was hoping that the Extension Host could still work even in this case with a setting to download extensions locally. Unfortunately this remote file is not working. Can you check to see if this was recently introduced? VS Code version: Code 1.96.3 (91fbdddc47bc9c09064bf7acf133d22631cbf083, 2025-01-09T18:14:09.060Z) OS version: Windows_NT x64 10.0.22631 Modes: Remote OS version: Linux x64 5.14.0-437.el9.x86_64 <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) Ultra 7 165U (14 x 2688)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|undefined| |Memory (System)|31.46GB (5.65GB free)| |Process Argv|--crash-reporter-id 7e9e6a5c-e81d-4f36-9073-13611d70a8a4| |Screen Reader|no| |VM|0%| |Item|Value| |---|---| |Remote|SSH: jumphost| |OS|Linux x64 5.14.0-437.el9.x86_64| |CPUs|Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz (2 x 0)| |Memory (System)|3.82GB (1.96GB free)| |VM|100%| </details><details><summary>Extensions (5)</summary> Extension|Author (truncated)|Version ---|---|--- catppuccin-vsc|Cat|3.16.0 remote-ssh|ms-|0.116.1 remote-ssh-edit|ms-|0.87.0 remote-explorer|ms-|0.4.3 ufuk-bakan-color-themes|Ufu|2.0.1 (3 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vspor879:30202332 vspor708:30202333 vspor363:30204092 vscod805:30301674 binariesv615:30325510 vsaa593:30376534 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 2i9eh265:30646982 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 h48ei257:31000450 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupyter:31046869 nativerepl2:31139839 pythonrstrctxt:31112756 nativeloc2:31192216 cf971741:31144450 iacca1:31171482 notype1:31157159 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 390bf810:31215807 6074i472:31201624 dwoutputs:31217127 ``` </details> <!-- generated by issue reporter -->
triage-needed,stale
low
Critical
2,796,242,535
material-ui
top-level-imports v5 changing color imports
### Steps to reproduce Steps: 1. Import grey from @mui/material/colors `import { grey } from @mui/material/colors;` 2. npx @mui/codemod@latest v5.0.0/top-level-imports <path> 3. The import will change to `import { grey } from '@mui/material';` ### Current behavior Steps: 1. Import grey from @mui/material/colors `import { grey } from @mui/material/colors;` 2. npx @mui/codemod@latest v5.0.0/top-level-imports <path> 3. The import will change to `import { grey } from '@mui/material';` ### Expected behavior `npx @mui/codemod@latest v5.0.0/top-level-imports <path>` should not change imports to those that are not valid ### Context _No response_ ### Your environment <details> <summary><code>npx @mui/envinfo</code></summary> ``` System: OS: macOS 15.1.1 Binaries: Node: 21.2.0 - ~/node_modules/.bin/node npm: 10.7.0 - ~/.nvm/versions/node/v18.20.3/bin/npm pnpm: Not Found Browsers: Chrome: 132.0.6834.83 Edge: 131.0.2903.147 Safari: 18.1.1 npmPackages: @emotion/react: ^11.10.4 => 11.14.0 @emotion/styled: ^11.10.4 => 11.14.0 @mui/base: 5.0.0-beta.66 @mui/core-downloads-tracker: 5.16.11 @mui/icons-material: ^5.8.4 => 5.16.11 @mui/lab: ^5.0.0-alpha.95 => 5.0.0-alpha.174 @mui/material: ^5.10.1 => 5.16.11 @mui/material-nextjs: ^5.15.6 => 5.16.8 @mui/private-theming: 5.16.8 @mui/styled-engine: 5.16.8 @mui/styles: ^5.9.3 => 5.16.11 @mui/system: 5.16.8 @mui/types: 7.2.19 @mui/utils: 5.16.8 @mui/x-data-grid: 6.20.4 @mui/x-data-grid-pro: ^6.0.4 => 6.20.4 @mui/x-date-pickers: 6.20.2 @mui/x-date-pickers-pro: ^6.0.0 => 6.20.2 @mui/x-license-pro: 6.10.2 @types/react: ^18.0.28 => 18.3.16 react: ^18.2.0 => 18.3.1 react-dom: ^18.2.0 => 18.3.1 typescript: ^4.9.5 => 4.9.5 ``` </details> **Search keywords**: codemod, top-level-import, color imports
bug 🐛,waiting for 👍,package: codemod,ready to take,v5.x
low
Minor
2,796,256,091
tauri
[bug] `cargo tauri dev` does not kill the application before recompiling Rust/config changes on Windows
### Describe the bug I used `cargo install [email protected]` with toolchain 1.84 and started migrating a v1 project to v2. Everything seems to be working, except when I run the dev server with `cargo tauri dev` with the file watcher enabled, it no longer kills the child process before trying to rerun `cargo build` when I change the Rust code or the Tauri config. I think it might be due to changing the child process runner to use `cargo run`, and that it's killing the `cargo.exe` process but not the `appname.exe` process spawned by that. ### Reproduction 1. Create a Tauri React app. 2. Install the `tauri-cli` binary locally instead of using the version downloaded by `yarn`. 3. Run `cargo tauri dev`. 4. Touch/save a change to `src-tauri/tauri.conf.json`. ### Expected behavior Along with killing the web dev server, and like on Unix-flavors, the app should be killed before rerunning `cargo build`. Otherwise, the file will still be in use and the dev server just errors out. ### Full `tauri info` output ```text [⚠] Environment - OS: Windows 10.0.26100 x86_64 (X64) ✔ WebView2: 131.0.2903.146 ✔ MSVC: Visual Studio Enterprise 2022 ✔ rustc: 1.84.0 [redacted] ✔ cargo: 1.84.0 ⚠ rustup: not installed! If you have rust installed some other way, we recommend uninstalling it then use rustup instead. Visit https://rustup.rs/ ⚠ Rust toolchain: couldn't be detected! Maybe you don't have rustup installed? if so, Visit https://rustup.rs/ - node: 22.5.1 - yarn: 1.22.19 - npm: 10.8.2 [-] Packages - tauri 🦀: 2.2.2 - tauri-build 🦀: 2.0.5 - wry 🦀: 0.48.1 - tao 🦀: 0.31.1 - tauri-cli 🦀: 2.2.4 - @tauri-apps/api : 2.2.0 - @tauri-apps/cli : 2.2.4 [-] Plugins - tauri-plugin-fs 🦀: 2.2.0 - @tauri-apps/plugin-fs : 2.2.0 [-] App - build-type: bundle - CSP: unset - frontendDist: ../dist - devUrl: http://localhost:1420/ ``` ### Stack trace ```text ``` ### Additional context We periodically rebuild the Rust toolchain internally, which is why there's no default `rustup` and why I redacted some of the `rustc` fingerprint in the `tauri info` output. It's also why I always `cargo install tauri-cli` locally and run everything with `cargo tauri` rather than `yarn tauri`.
type: bug,status: needs triage
low
Critical
2,796,258,184
PowerToys
Task Scheduler fails to execute autorun
### Microsoft PowerToys version 0.87.1 ### Installation method PowerToys auto-update ### Running as admin Yes ### Area(s) with issue? General ### Steps to reproduce This is one of the "Autorun not working at startup" issues. The issue has only shown up within the past month after using PowerToys the last 4 years. I have been perusing the previous threads without much success, so I am sharing what I have attempted so far. - Toggling autorun on/off in the general settings UI. - Tweaking Task Scheduler's startup condition (e.g., launching at startup regardless of log on behaviour) Now Task Scheduler is being odd here. It claims to have successfully executed the scheduled task, but the Last Run Result displays the 0x41303 string (Not Running). Running the task manually does start PowerToys. ![Image](https://github.com/user-attachments/assets/df016e27-af98-45c1-bb15-76d0575b8331) ![Image](https://github.com/user-attachments/assets/673e9f60-664c-4ef3-ab65-318892ac3abd) ### ✔️ Expected Behavior PowerToys launching on startup ### ❌ Actual Behavior PowerToys not launching on startup ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,796,316,504
ui
[bug]: CLI not respecting settings in components.json
### Describe the bug ```json { "$schema": "https://ui.shadcn.com/schema.json", "style": "new-york", "rsc": false, "tsx": true, "tailwind": { "config": "tailwind.config.ts", "css": "app/app.css", "baseColor": "zinc", "cssVariables": true, "prefix": "", "tsx": true }, "aliases": { "components": "~/app/components", "utils": "~/app/utilities", "ui": "~/app/components/ui", "lib": "~/app/utilities", "hooks": "~/app/hooks" }, "iconLibrary": "lucide" } ``` I have this for my components.json and whenever I install a new component the import for `cn` is always `import { cn } from "~/lib/utils"` ### Affected component/components CLI ### How to reproduce 1. Install shadcn 2. Modify `components.json` with a new `utils` alias 3. Install any component (i.e.: `pnpm dlx shadcn@latest add checkbox`) 4. Open the checkbox and see that the config wasn't respected ### Codesandbox/StackBlitz link _No response_ ### Logs ```bash ``` ### System Info ```bash MacOS, Vite, React Router 7 ``` ### Before submitting - [x] I've made research efforts and searched the documentation - [x] I've searched for existing issues
bug
low
Critical
2,796,341,555
PowerToys
[Github] PRs missing build CI checks
### Repository Branch main ### ❌ Actual Behavior The build ci checks aren't running anymore after updating PRs against current main. Why? xref: #33703 ![Image](https://github.com/user-attachments/assets/408b3350-22af-45fe-8e6c-d20a5310bcda) ### ✔️ Expected Behavior Build CI checks are executed after pushing changes.
Issue-Bug,Priority-0,Area-Build,Needs-Team-Response,Status-Reproducible
low
Minor
2,796,343,916
langchain
docs: Replace `initialize_agent` use with langgraph.prebuilt.create_react_agent
### Privileged issue - [x] I am a LangChain maintainer, or was asked directly by a LangChain maintainer to create an issue here. ### Issue Content Currently there's a bunch of tools and other integrations that use `initialize_agent` in their docs page (search langchain/docs for "initialize_agent") This function has been deprecated since 0.1 and should be replaced with basic langgraph usage as a demo Deprecated function: https://python.langchain.com/api_reference/langchain/agents/langchain.agents.initialize.initialize_agent.html Example page using create_react_agent in an example: https://python.langchain.com/docs/integrations/tools/gmail/#use-within-an-agent This issue doesn't need to be resolved by a single PR and can be tackled incrementally! Just tag this issue for tracking purposes :)
documentation,help wanted,good first issue,🤖:docs
low
Minor
2,796,350,085
rust
Tracking issue for release notes of #135656: Add `-C hint-mostly-unused` to tell rustc that most of a crate will go unused
This issue tracks the release notes text for #135656. ### Steps - [ ] Proposed text is drafted by PR author (or team) making the noteworthy change. - [ ] Issue is nominated for release team review of clarity for wider audience. - [ ] Release team includes text in release notes/blog posts. ### Release notes text The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing). ````markdown # Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...) - [Add `-C hint-mostly-unused` to tell rustc that most of a crate will go unused](https://github.com/rust-lang/rust/pull/135656) ```` > [!TIP] > Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use. > The category will be de-duplicated with all the other ones by the release team. > > *More than one section can be included if needed.* ### Release blog section If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section. *Otherwise leave it empty.* ````markdown ```` cc @joshtriplett, @fee1-dead -- origin issue/PR authors and assignees for starting to draft text
T-compiler,relnotes,relnotes-tracking-issue
low
Minor
2,796,353,567
langchain
Update default values of minimax_api_host and model in MiniMaxChat __init__
### Checked other resources - [x] I added a very descriptive title to this issue. - [x] I searched the LangChain documentation with the integrated search. - [x] I used the GitHub search to find a similar question and didn't find it. - [x] I am sure that this is a bug in LangChain rather than my code. - [x] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code The following code: ```python from langchain_community.chat_models import MiniMaxChat api_key = "YOUR_API_KEY" chat = MiniMaxChat(api_key=api_key) messages = [ ("human", "Hello, how are you?"), ] resp = chat.invoke(messages) print(resp) ``` ### Error Message and Stack Trace (if applicable) ``` Traceback (most recent call last): File "/home/work/langchain/bug_report.py", line 13, in <module> resp = chat.invoke(messages) File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 286, in invoke self.generate_prompt( File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 786, in generate_prompt return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 643, in generate raise e File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 633, in generate self._generate_with_cache( File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 851, in _generate_with_cache result = self._generate( File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_community/chat_models/minimax.py", line 515, in _generate return self._create_chat_result(response.json()) File "/home/work/langchain/venv/lib/python3.10/site-packages/langchain_community/chat_models/minimax.py", line 429, in _create_chat_result for res in response["choices"]: KeyError: 'choices' ``` ### Description * I am trying to run `MiniMaxChat`using default settings * The underlying problem is that the default values for parameters do not work - `minimax_api_host` is `https://api.minimax.chat/v1/text/chatcompletion_v2`, whereas it should be `https://api.minimaxi.chat/v1/text/chatcompletion_v2` (note the `i` -> `minimax` -> `minimaxi`) - `model` is `abab6.5-chat`, whereas it should be `abab6.5s-chat` (note the `s`) - Consequently, the received response is not a succesful one, so the `choices` key is not present in it * Using overriden parameters `MiniMaxChat(api_key=api_key, model="abab6.5s-chat", minimax_api_host="https://api.minimaxi.chat/v1/text/chatcompletion_v2")` yields successful result * Additionally there is a problem with the code itself at the point of the raised exception. It's difficult to comprehend for a user why the `KeyError` was raised - but that is for another issue probably. - In my opinion, it would be better to fail early with an informative error message - invalid API key, insufficient balance, invalid model name etc. ### System Info `python -m langchain_core.sys_info` yields: ``` System Information ------------------ > OS: Linux > OS Version: #52~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Mon Dec 9 15:00:52 UTC 2 > Python Version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0] Package Information ------------------- > langchain_core: 0.3.30 > langchain: 0.3.14 > langchain_community: 0.3.14 > langsmith: 0.2.11 > langchain_minimax: Installed. No version info available. > langchain_openai: 0.3.0 > langchain_text_splitters: 0.3.5 Optional packages not installed ------------------------------- > langserve Other Dependencies ------------------ > aiohttp: 3.11.11 > async-timeout: 4.0.3 > dataclasses-json: 0.6.7 > httpx: 0.28.1 > httpx-sse: 0.4.0 > jsonpatch: 1.33 > langsmith-pyo3: Installed. No version info available. > numpy: 1.26.4 > openai: 1.59.8 > orjson: 3.10.14 > packaging: 24.2 > pydantic: 2.10.5 > pydantic-settings: 2.7.1 > PyYAML: 6.0.2 > requests: 2.32.3 > requests-toolbelt: 1.0.0 > SQLAlchemy: 2.0.37 > tenacity: 9.0.0 > tiktoken: 0.8.0 > typing-extensions: 4.12.2 > zstandard: Installed. No version info available. ```
🤖:bug
low
Critical
2,796,356,520
godot
[Godot 4.4 Beta1] Big TileMapLayer's tiles disappear and flicker when used inside a CanvasGroup and zooming out
### Tested versions Reproducible in: 4.4 Beta1, 4.4 dev7. Not reproducible in: 4.1.4 stable. ### System information Godot v4.4.beta1 - Windows 10 (build 19045) - Multi-window, 1 monitor - Vulkan (Forward+) - integrated AMD Radeon(TM) Vega 8 Graphics (Advanced Micro Devices, Inc.; 27.20.1028.1) - AMD Ryzen 5 3500U with Radeon Vega Mobile Gfx (8 threads) ### Issue description I was switching a project from Godot 4.1.4 to 4.4 when I noticed the minimap I made for it disappeared, it was also flickering in and out every time the character moved. The setup for it was really simple, so i made a test project to confirm if it also happened from a blank project and it did. (I made a shader to make this easier to view in the reproduction project and works without it, the TileMapLayer outside the CanvasGroup has black tiles and the one inside the CanvasGroup has white tiles) ![Image](https://github.com/user-attachments/assets/e79b835f-e893-4896-bd01-a922646bbbcf) Zooming out ![Image](https://github.com/user-attachments/assets/99b015c2-fc5e-4504-916a-33435a0c7ecc) This works both in the editor and when running the project. As to why I am using a CanvasGroup with a TileMapLayer this big: I made a shader in 4.1.4 using the texture CanvasGroup gives, it allowed me to make effects I really liked, but I can't do the same effects without it. Changing the fit margin/clear margin/mipmaps in the CanvasGroup doesn't seem to do anything even when set to ridiculous values. ### Steps to reproduce -Make a scene with a CanvasGroup. -Add a TileMapLayer as a child of the CanvasGroup. -Add a TileSet to the TileMapLayer and put any texture to it to use as tiles. -Use the Rect Tool to fill a very large area of the world with tiles. -Zoom the camera out. ### Minimal reproduction project (MRP) [canvasgroupbigtilemap.zip](https://github.com/user-attachments/files/18460023/canvasgroupbigtilemap.zip)
bug,topic:editor,needs testing,topic:2d
low
Minor
2,796,381,668
go
cmd/go: `all` pattern in per package flags is inaccurate in module mode
`all` has a slightly different meaning in module mode than it does in `GOPATH` mode. In `GOPATH` mode, it's pretty much everything: > In GOPATH mode, "all" expands to all packages found in all the GOPATH trees. But in module mode, "all" > expands to the set of "packages in the main module (or workspace modules) and their dependencies, including dependencies needed by tests of any of those" according to `go help packages`. We don't seem to handle this correctly when processing "all" in `-coverpkg` and the per package flags `-asmflags`, `-gccgoflags`, `-gcflags`, and `-ldflags`. The cmd/go/internal/load.MatchPackage function just matches every package for all.
NeedsInvestigation,GoCommand
low
Minor
2,796,383,244
pytorch
Obey sm_carveout (limit on number of SMs) in inductor persistent kernel
### 🚀 The feature, motivation and pitch See https://github.com/pytorch/pytorch/pull/144974#issuecomment-2599011250 cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @aakhundov @eellison @lw ### Alternatives _No response_ ### Additional context _No response_
triaged,oncall: pt2,module: inductor
low
Minor
2,796,397,162
rust
E0106 "missing lifetime specifier" should be omitted when the signature is incorrect in an `impl` (E0186)
### Code ```Rust trait Foo { fn foo(&self) -> &str; } struct Example; impl Foo for Example { fn foo() -> &str { "example" } } ``` ### Current output ```Shell error[E0106]: missing lifetime specifier --> src/lib.rs:7:17 | 7 | fn foo() -> &str { | ^ expected named lifetime parameter | = help: this function's return type contains a borrowed value, but there is no value for it to be borrowed from help: consider using the `'static` lifetime, but this is uncommon unless you're returning a borrowed value from a `const` or a `static` | 7 | fn foo() -> &'static str { | +++++++ help: instead, you are more likely to want to return an owned value | 7 | fn foo() -> String { | ~~~~~~ error[E0186]: method `foo` has a `&self` declaration in the trait, but not in the impl --> src/lib.rs:7:5 | 2 | fn foo(&self) -> &str; | ---------------------- `&self` used in trait ... 7 | fn foo() -> &str { | ^^^^^^^^^^^^^^^^ expected `&self` in impl ``` ### Desired output ```Shell error[E0186]: method `foo` has a `&self` declaration in the trait, but not in the impl --> src/lib.rs:7:5 | 2 | fn foo(&self) -> &str; | ---------------------- `&self` used in trait ... 7 | fn foo() -> &str { | ^^^^^^^^^^^^^^^^ expected `&self` in impl ``` ### Rationale and extra context When the function is a trait associated function, the signature must match (or refine) the trait’s, so E0106 and any other errors that would propose changes to the signature are misleading or at least not nearly as relevant as E0186 (or E0050 or E0053 or any other such mismatch). Therefore, they should be hidden. This is particularly important for situations involving a missing `self` parameter, because (in my personal experience) it’s easy to forget to write `self` in the parameter list. ### Rust Version ```Shell 1.84.0 ```
A-diagnostics,T-compiler
low
Critical
2,796,412,976
react-native
Zooming in ScrollView causes incorrect offset of content
### Description When using the zoom and pinchGestureEnabled on a ScrollView, the content shifts to the side after re-rendering the component when zoomed in. ### Steps to reproduce 1) Toggle the ScrollView on 2) Zoom in 3) Toggle the ScrollView off 4) Toggle the ScrollView on ### React Native Version 0.76.6 ### Affected Platforms Runtime - iOS ### Output of `npx react-native info` ```text System: OS: Windows 11 10.0.22631 CPU: (12) x64 Intel(R) Core(TM) i7-10750H CPU @ 2.60GHz Memory: 3.26 GB / 15.78 GB Binaries: Node: version: 22.8.0 path: C:\Program Files\nodejs\node.EXE Yarn: version: 1.22.22 path: C:\Program Files\nodejs\yarn.CMD npm: version: 10.8.2 path: C:\Program Files\nodejs\npm.CMD Watchman: Not Found SDKs: Android SDK: Not Found Windows SDK: Not Found IDEs: Android Studio: Version 2021.1.0.0 AI-211.7628.21.2111.8193401 Visual Studio: - 17.3.32804.467 (Visual Studio Community 2022) Languages: Java: version: 17.0.8 path: /c/Program Files/Common Files/Oracle/Java/javapath/javac Ruby: Not Found npmPackages: "@react-native-community/cli": installed: 15.1.3 wanted: ^15.1.3 react: installed: 18.3.1 wanted: 18.3.1 react-native: installed: 0.76.6 wanted: 0.76.6 react-native-windows: Not Found npmGlobalPackages: "*react-native*": Not Found Android: hermesEnabled: Not found newArchEnabled: Not found iOS: hermesEnabled: Not found newArchEnabled: Not found ``` ### Stacktrace or Logs ```text Visual bug ``` ### Reproducer https://snack.expo.dev/@bram-dc/scrollview-zoom-offset ### Screenshots and Videos https://github.com/user-attachments/assets/7eaed09f-1b25-4cb1-bb7d-379bf86b29d4
Issue: Author Provided Repro,Component: ScrollView,Needs: Version Info
low
Critical
2,796,457,228
go
x/build/cmd/relui: add support for making release announcement posts on Bluesky
During Go releases, relui takes on posting on 2 social sites using their APIs (#47403 and #65046). By now we have an account on one more social site, Bluesky, where we've also started to post links to Go release announcements (for example, https://bsky.app/profile/golang.org/post/3lfvcgeox7c2c). This is the tracking issue for automating Go release announcement posts on [Bluesky](https://docs.bsky.app/). CC @golang/release, @mknyszek, @fflewddur.
Builders,NeedsInvestigation,Automation
low
Minor
2,796,462,469
flutter
[flutter_test] gestures can leak from test to test
Steps to reproduce: * Check out this revision of Flutter: https://github.com/flutter/flutter/commit/bb2d34126cc8161dbe4a1bf23c925e48b732f670 * `cd packages/flutter` * `flutter test --reporter=expanded --test-randomize-ordering-seed=20250118 --fatal-warnings --track-widget-creation test/material/reorderable_list_test.dart` * Observe the test passing. * Now change the seed: `flutter test --reporter=expanded --test-randomize-ordering-seed=20250117 --fatal-warnings --track-widget-creation test/material/reorderable_list_test.dart` * Observe the test failing. The reason it's failing is because a test in `reorderable_list_test.dart` created a `TestGesture` but never finalized it by cancelling it or by calling `drag.up()`. Because tests that are sensitive to dangling gestures could be anywhere, this issue did not immediately reproduce until, by chance, we rolled a seed that sorts tests such that the sensitive test was run immediately after the test that leaves the gesture hanging (which took about 1 month). This was fixed in https://github.com/flutter/flutter/pull/161836. However, the issue should have been caught immediately when `reorderable_list_test.dart` was added for the first time. `flutter_test` could validate that when a test ends, there are not dangling gestures and `VelocityTrackers` left.
a: tests,framework,P1,team-framework,triaged-framework
medium
Minor
2,796,500,511
vscode
Line normalization in `Terminal.sendText()` (IPython)
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> <details> <summary>Show System Information</summary> - VSCode: Version: 1.96.4 (user setup) Commit: cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba Date: 2025-01-16T00:16:19.038Z Electron: 32.2.6 ElectronBuildId: 10629634 Chromium: 128.0.6613.186 Node.js: 20.18.1 V8: 12.8.374.38-electron.0 OS: Windows_NT x64 10.0.19045 - I'm using Windows 10, Version 22H2 with WSL2 (Ubuntu 22.04) </details> --- ### 🎈 `Terminal.sendText()` The `Terminal.sendText()` function currently normalizes line endings as follows: https://github.com/microsoft/vscode/blob/0d57cf5304672f3cd0c2776aaec1e5dedfead053/src/vs/workbench/contrib/terminal/browser/terminalInstance.ts#L1271-L1275 With regards to cross-platform operability, I am wondering **why we don't append any `\n` here and instead just a single `\r`**? E.g. the default line separation on Windows is `\r\n` and not `\r`. Git Blame: The mentioned code was first proposed in [this comment](https://github.com/microsoft/vscode/issues/22887#issuecomment-296505048) and then introduced via [this commit](https://github.com/microsoft/vscode/commit/d01c71579d6690e7a7c8311f8477275d8b9b0203). During this, the `os.EOL` was replaced by `\r` and I don't see the reason to do so. --- ### 🎈 Possible symptoms I suspect this normalization to have caused [this problem](https://github.com/ipython/ipython/issues/13054#issuecomment-2599055070) for us in the role of VSCode extension authors. For that, assume that the extension user has opened a VSCode-integrated terminal and started an IPython shell inside it (IPython: interactive Python shell, feels like a REPL for Python). Assume further that the user has defined a Python method `def hello_world()` inside that IPython shell. In our extension, we send the string `hello_world()` to that terminal and expect that the `hello_world()` method is run. What we actually find, is that IPython enters the multi-line mode, so we see the following: ```py In [2]: hello_world() ..: ``` This is only the case on Windows. On macOS and Unix it works fine. I'm not entirely sure whether this problem is due to this normalization or due to some IPython behavior. [This issue](https://github.com/ipython/ipython/issues/13054) is already open for it. Furthermore, because of this issue, the VSCode Python extension had to add a manual timeout, which might fail on slower machines. See their `sendText()` method [here](https://github.com/saponas/vscode-python/blob/1597f84ec29f320fa707355d340282994417f407/src/client/common/terminal/service.ts#L48-L56) and my [summary here](https://github.com/ipython/ipython/issues/13054#issuecomment-2598897658). ```ts this.terminal!.sendText(`${text}`); // send carriage return to make code execute await new Promise(resolve => setTimeout(resolve, 100)); this.terminal.sendText('\u000D', false); ```
under-discussion
low
Major
2,796,502,077
flutter
Engine git-hooks are executed even for framework-only PRs
```sh matanl@matanl-macbookpro4 flutter % git push -f The clang-tidy check is disabled. To enable set the environment variable PRE_PUSH_CLANG_TIDY to any value. Starting formatting checks. ``` Not only is this taking forever (I believe unrelated), there is nothing it could be checking in this PR.
team-infra,P2,monorepo
low
Minor
2,796,512,009
ant-design
Buttons with variant "link" or "text" use the `info` color when the color is set to `primary`
### Reproduction link [![Edit on StackBlitz](https://developer.stackblitz.com/img/open_in_stackblitz.svg)](https://stackblitz.com/edit/react-t3ejzn?file=demo.tsx) ### Steps to reproduce - Customize primary color and info color - Use a button and pass the props color="primary" variant="text" ### What is expected? The text color should be the primary color ### What is actually happening? The text color is the info color | Environment | Info | | --- | --- | | antd | 5.23.1 | | React | doesn't matter | | System | doesn't matter | | Browser | doesn't matter | --- So it is basically the same issue as https://github.com/ant-design/ant-design/issues/51343. I know you closed it but I really want you to reconsider it. With the new colors of buttons (which released in version `5.23.0`) the `text` variant is working great. I think that if you explicitly set the button color to `primary` it must have a primary color. <!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
unconfirmed
low
Minor
2,796,519,893
flutter
Use `DEPS` to bootstrap the GitHub Actions Dart SDK
Some pondering on if initializing the Dart SDK can be vastly simplified, potentially also removing the need for the engine to create a copy of the Dart SDK as part of vended artifacts. One of the reasons this _could_ be beneficial is we'd like to explore using GitHub actions in the Flutter repo. A GitHub action would _not_ wait for the engine to be built, therefore it would not have access to the engine binaries for the current SHA, and for the action to be written in Dart (instead of say, Python or Bash), it would need a version of the Dart SDK from somewhere - and ideally that would be the same version as pinned in Flutter for compatibility reasons. In https://github.com/flutter/flutter/pull/161788 I experiment with an override, `FLUTTER_ENGINE_VERSION`, that allows the GitHub action to use the Dart SDK from the base-ref (which merged commit the PR was branched off originally). This will work _until the Dart SDK revs_, in which case now we'll be using an older Dart SDK to evaluate the Flutter repo. Potential issues: what if we wanted to run `dart format` on a GitHub action? We'd be using an older version of Dart format than the one potentially being changed/updated in `DEPS`. The proposal below is an _idea_ of what we could do, but we could imagine other ideas/overrides, for example, we could just have the GitHub _action_ self-bootstrap the Dart SDK from `DEPS` and _not_ change how `bin/dart` works at all. --- Today, `bin/dart` (part of the tools shipped in the Flutter SDK) is a wrapper script that self-bootstraps: https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/dart#L46-L59 It, if necessary, downloads the Dart SDK, by writing a `engine.version` file: https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/internal/update_engine_version.sh#L41 ... and then using that value in a later shell script to download a Dart SDK: https://github.com/flutter/flutter/blob/d59d02c014428ac71eb347165d4616043354b7f4/bin/internal/shared.sh#L147 https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/bin/internal/update_dart_sdk.sh#L126-L161 Lots of this logic is due to the fact that the engine and framework repos were separate, there was no (canonical) way to identify which Dart SDK to use, but now there is a single canonical number that is used: https://github.com/flutter/flutter/blob/78c62b00cc58b785ae1d6fca44291634b0c8b3c0/DEPS#L59
team-infra,P3,c: tech-debt,monorepo
low
Major
2,796,520,756
angular
⚠️ Local development bugs with HMR
### Which @angular/* package(s) are the source of the bug? compiler-cli ### Is this a regression? Yes ### Description We're aware of reports of local development bugs arising from Angular template HMR in 19.1.x, which has been enabled by default. List of known issues: * [Fixed in 19.1.2] ~~#59581~~ * [Fixed in 19.1.2] ~~#59592~~ * [Fixed in 19.1.2] ~~angular/angular-cli#29248~~ * [Fixed in 19.1.3(CLI)] ~~#59605~~ * [Fixed in 19.1.4(CLI)] ~~angular/angular-cli#29424~~ * [Fixed in 19.1.3] ~~#59632~~ * [Fixed in 19.1.3] ~~#59634~~ * [Fixed in 19.1.3] ~~#59639~~ * [Fixed in 19.1.3] ~~#59640~~ We're actively investigating and fixing these issues to create a smooth HMR experience out of the box. Production builds and releases are not affected. If you discover any additional HMR challenges which don't appear to be related to the linked issues above, verify that you are on a recent version (19.1.3+), and please [file a new issue](http://github.com/angular/angular/issues/) and link it here so we can triage and investigate. ### Workarounds In the meantime, if you encounter any problems relating to HMR, you can work around them in three ways: 1. Refresh the page \- This loads a fresh copy of your application with up to date code. 2. `ng serve --no-hmr` \- This disables HMR for your local devserver and automatically refreshes the page on any code changes. 3. Disable HMR for your project \- Configuring `angular.json` will disable HMR for all `ng serve` invocations by default. * This is useful for a project with a large number of developers without needing to communicate these workarounds to all contributors. * Remember to re-enable once the relevant issues are fixed\! ```json { "projects": { "my-project": { "architect": { "serve": { "options": { "hmr": false } } } } } } ``` Thank you for sharing any HMR issues and helping improve Angular one bugfix at a time. We're committed to fully stabilizing this feature and excited to bring improved build performance to everyone.
area: core,core: hot module replacement (HMR)
medium
Critical
2,796,528,295
flutter
A label to track approval of `c: proposal` issues?
For example, in https://github.com/flutter/flutter/issues/161843, I propose making a change. It would be nice for us to mark proposals as approved, even if it's just a GitHub label, so that folks can know "this is a proposal, and the domain owner(s) agree with it being implemented" versus "this is a proposal, and we haven't reviewed/rejected/approved it yet".
team,will need additional triage,c: proposal,fyi-infra
low
Minor
2,796,545,097
rust
Delay stabilizing async closures to consider if they should return `impl IntoFuture` instead of `impl Future`
Async closures are currently in beta for stabilization in 1.85. I'm very excited to see this feature stabilized! That said, I would like the Lang team to consider delaying their stabilization to consider a new design detail. Delaying stabilization would amount to putting a PR in nightly that undoes the stabilization, and then doing a beta backport. Note that this doesn't necessarily mean the lang team would be committing to making this change, but if the lang team feels there's a non-trivial chance they would prefer the `IntoFuture` semantics, then we should delay stabilization so there is time to consider this. I talked with @tmandry and @traviscross about this and they did not remember discussing this particular detail during the stabilization process. If we are mistaken and this was discussed, then we should let the decision stand and not change anything. # Why should T-lang consider this change? @yoshuawuyts recently published a [blog post about generators and auto trait leakage](https://blog.yoshuawuyts.com/gen-auto-trait-problem/), which argues that `gen {}` should return `impl IntoIterator` instead of `impl Iterator`. The key example in the post is (adapted slightly): ```rust let iter = gen { let rc = Rc::new(...); yield 12u32; rc.do_something(); }; spawn(|| { for num in iter { println!("{num}"); } }).unwrap(); ``` If `gen {}` returns `impl Iterator`, as it does on nightly today, this example fails to compile because the `Rc` is not `Send`. However, this is overly restrictive because the `Rc` is not created until the spawned thread calls `next()` the first time. Instead, having `gen {}` return `impl IntoIterator` allows this to work, because the `IntoIterator` part could be `Send`, and once it becomes an `Iterator` it would no longer be `Send`. This has a similar shape to annoyances we've seen with async Rust, where it's not uncommon for people to want to write futures that are `Send` until they are polled for the first time. If we had made `async {}` blocks return `impl IntoFuture` instead, this would be straightforward to do for the same reasons. So then the next logical question is what should generator closures or async closures do? For much the same reason, it would be nice for async closures to return futures that are `Send` until first polled. At this point, changing the return type for `async {}` would require an edition change. However, we have a brief window where we can still make this change for async closures. # Why should T-lang keep the status quo? A lot of care has gone into the current design by @compiler-errors and others to come up with something that balances a lot of competing goals. For example, with the current design, the following example works, but it would break with this proposed change: ```rust fn foo<F, U>(_f: F) where F: Fn() -> U, U: Future {} // can be invoked with legacy, fake async closures or new, real async closures foo(|| async {}); foo(async || {}); ``` This is a desirable property, since a lot of existing ecosystem code is written using `Fn() -> impl Future` as an approximation of an async closure. If async closures returned `IntoFuture` then they would not work until existing code is updated to support async closures natively. I bring this up mainly to show that it's not obvious that we actually should make this change. There are many factors in tension and there's a good chance we already have the optimal balance. We'd want to thoroughly consider any changes, so this issue is mostly about whether we want to take the time to consider those changes. # Mitigating Factors This would still be changeable over an edition if we were to decide it's worthwhile but don't want to churn things now. If we were to make this change, we'd likely want to change the behavior of `async {}`. Doing both over an edition at the same time would keep the overall language more consistent.
T-lang,C-discussion,I-lang-nominated
medium
Critical
2,796,550,635
godot
Editor crash while refactoring scripts
### Tested versions - Occurred in v4.4.dev7.official [46c8f8c5c] ### System information Godot v4.4.dev7 - Windows 11 (build 22631) - Multi-window, 1 monitor - OpenGL 3 (Compatibility) - NVIDIA GeForce RTX 3080 Ti (NVIDIA; 32.0.15.6636) - AMD Ryzen 9 5900X 12-Core Processor (24 threads) ### Issue description I experienced two crashes while performing some script/class refactoring in the Godot editor. The exact repro steps are unknown, however, I have two crash dumps from Windows. Both are double frees, so my best guess is an issue with saving and compiling scripts sharing resources improperly. [Godot_v4.4-dev7_win64.exe.10648.dmp](https://github.com/user-attachments/files/18462120/Godot_v4.4-dev7_win64.exe.10648.dmp) [Godot_v4.4-dev7_win64.exe.25672.dmp](https://github.com/user-attachments/files/18462119/Godot_v4.4-dev7_win64.exe.25672.dmp) ### Steps to reproduce Some combination of the following actions in quick succession in an unknown order: (e.g. when duplicating `wall_*.gd` for a new type) - Ctrl+D (duplicate) a script with a `class_name` - Double click to open the script - Ctrl+S to save the script - Updating the `class_name {name}` name ### Minimal reproduction project (MRP) Sample project to bootstrap reproduction [poc.zip](https://github.com/user-attachments/files/18462323/poc.zip)
bug,topic:gdscript,topic:editor,crash
low
Critical
2,796,555,929
rust
LLVM assertion failure on Wasm exceptions
### Code ```Rust fn main() {} ``` :/ ### Meta I'm building rustc on the current master (bcd0683e5dce1945b5d940714742e7502883bb5c) with LLVM assertions enabled. I'm compiling the program above like this: ``` RUSTFLAGS="-C panic=unwind" cargo +stage1 build --target wasm32-wasip1 -Z build-std ``` I'm using the 23.0 WASI sysroot, downloaded from their GitHub releases, if that matters. ### Error output ``` rustc: /checkout/src/llvm-project/llvm/lib/CodeGen/SelectionDAG/SelectionDAGISel.cpp:1378: void mapWasmLandingPadIndex(MachineBasicBlock *, const CatchPadInst *): Assertion `IntrFound && "wasm.landingpad.index intrinsic not found!"' failed. ``` like, everywhere, on all sorts of standard crates like `libc` and `unwind`. Disabling assertions hides the bug, and user code works fine. This also does not reproduce on rustup's nightly, probably for the same reason. I have no idea, really. Please tell me I'm doing something wrong. For context, this seems more important because Emscripten is supposed to switch to Wasm exceptions soon-ish, and I imagine the same problem occurs there (but I haven't checked). Unlike wasm32-wasip1, unwinding is enabled there by default. @rustbot label +A-LLVM
I-crash,A-LLVM,T-compiler,O-wasm,C-bug,O-wasi,O-emscripten
medium
Critical
2,796,566,382
ollama
HSA_OVERRIDE_GFX_VERSION_0 while running on only one GPU
### What is the issue? I am running `ollama:rocm` in a docker container on Ubuntu 24.04. My GPU is an RX 6600 (`gfx1032`). Everything works fine when I run `ollama` using ```bash docker run -d \ --device /dev/kfd \ --device /dev/dri \ -v ollama:/root/.ollama \ -p 11434:11434 \ --restart unless-stopped \ --env HSA_OVERRIDE_GFX_VERSION="10.3.0" \ --env OLLAMA_KEEP_ALIVE="-1" \ --name ollama-rocm \ ollama/ollama:rocm ``` ``` time=2025-01-18T01:24:54.470Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.5-0-g32bd37a-dirty)" time=2025-01-18T01:24:54.470Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu_avx2 rocm_avx cpu cpu_avx]" time=2025-01-18T01:24:54.470Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs" time=2025-01-18T01:24:54.472Z level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory" time=2025-01-18T01:24:54.472Z level=INFO source=amd_linux.go:391 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION=10.3.0 time=2025-01-18T01:24:54.472Z level=INFO source=types.go:131 msg="inference compute" id=0 library=rocm variant="" compute=gfx1032 driver=0.0 name=1002:73ff total="8.0 GiB" available="8.0 GiB" ``` However, when I run using ```bash docker run -d \ --device /dev/kfd \ --device /dev/dri \ -v ollama:/root/.ollama \ -p 11434:11434 \ --restart unless-stopped \ --env HSA_OVERRIDE_GFX_VERSION_0="10.3.0" \ --env OLLAMA_KEEP_ALIVE="-1" \ --name ollama-rocm \ ollama/ollama:rocm ``` `ollama` runs only on CPU: ``` time=2025-01-18T01:25:58.373Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.5-0-g32bd37a-dirty)" time=2025-01-18T01:25:58.373Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu_avx2 rocm_avx cpu cpu_avx]" time=2025-01-18T01:25:58.373Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs" time=2025-01-18T01:25:58.375Z level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory" time=2025-01-18T01:25:58.378Z level=WARN source=amd_linux.go:378 msg="amdgpu is not supported (supported types:[gfx1030 gfx1100 gfx1101 gfx1102 gfx900 gfx906 gfx908 gfx90a gfx940 gfx941 gfx942])" gpu_type=gfx1032 gpu=0 library=/usr/lib/ollama time=2025-01-18T01:25:58.378Z level=WARN source=amd_linux.go:385 msg="See https://github.com/ollama/ollama/blob/main/docs/gpu.md#overrides for HSA_OVERRIDE_GFX_VERSION usage" time=2025-01-18T01:25:58.378Z level=INFO source=amd_linux.go:404 msg="no compatible amdgpu devices detected" time=2025-01-18T01:25:58.378Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered" time=2025-01-18T01:25:58.378Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="62.7 GiB" available="61.2 GiB" ``` I am trying this because eventually I want to get newer AMD GPUs and use them concurrently for `ollama`. Is this not supported when running `ollama` in `docker`? If I want to use this RX 6600 and an RX 7800 in the same system, how should I do it with `docker`? ### OS Linux ### GPU AMD ### CPU Intel ### Ollama version 0.5.5-0-g32bd37a-dirty
bug
low
Critical
2,796,575,411
ollama
ollama/ollama:rocm not detecting AMD GPU being passed in
### What is the issue? # Overview I've been trying to add ollama/openwebui to my home lab and I've run across an issue I cannot figure out. ROCM has finally been released last month for my linux distro and I've got it installed and can run (on gpu) ollama on my host machine. When I try to transition to docker however, I am unable to get it to work. I will provide what I have but please ask any clarifying questions. ## Components Host: Debian 6.1.124-1 CPU: I7-12700K GPU: Radeon 6900 XT # Docker compose file ``` services: ollama: image: ollama/ollama:rocm container_name: ollama hostname: ollama privileged: true volumes: - ollama:/root/.ollama environment: HSA_OVERRIDE_GFX_VERSION: 10.3.0 AMD_SERIALIZE_KERNEL: 3 HIP_VISIBLE_DEVICES: 0 OLLAMA_DEBUG: 1 AMD_LOG_LEVEL: 3 # ports: # - :11434 restart: unless-stopped networks: - npm devices: - /dev/kfd - /dev/dri group_add: - 105 - 44 security_opt: - seccomp:unconfined ``` # Logs & Outputs No errors that I can really tell. Here are the logs on startup in docker ``` 2025-01-17 20:40:33 2025/01/18 01:40:33 routes.go:1187: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES:0 HSA_OVERRIDE_GFX_VERSION:10.3.0 HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:true OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.654Z level=INFO source=images.go:432 msg="total blobs: 0" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=images.go:439 msg="total unused blobs removed: 0" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.7-0-ga420a45-dirty)" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:80 msg="runners located" dir=/usr/lib/ollama/runners 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx/ollama_llama_server 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx2/ollama_llama_server 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/rocm_avx/ollama_llama_server 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cpu cpu_avx cpu_avx2 rocm_avx]" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=routes.go:1268 msg="Override detection logic by setting OLLAMA_LLM_LIBRARY" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.655Z level=DEBUG source=sched.go:105 msg="starting llm scheduler" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:99 msg="searching for GPU discovery libraries for NVIDIA" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcuda.so* 2025-01-17 20:40:33 time=2025-01-18T01:40:33.656Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/lib/ollama/libcuda.so* /libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths=[] 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcudart.so* 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /libcudart.so* /usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths=[] 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=DEBUG source=amd_linux.go:421 msg="amdgpu driver not detected /sys/module/amdgpu" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered" 2025-01-17 20:40:33 time=2025-01-18T01:40:33.657Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="50.7 GiB" available="35.1 GiB" ``` ```ls -ld /sys/module/amdgpu``` is display ```drwxr-xr-x 7 root root 0 Jan 17 12:23 /sys/module/amdgpu``` ### OS Linux, Docker ### GPU AMD ### CPU Intel ### Ollama version 0.5.7-0-ga420a45-dirty
bug
low
Critical
2,796,582,590
tauri
linux development often show blank screen.
when modify some code , sometimes the `webview` component show nothing. I need restart `pnpm run dev` and restart the rust `backend` to make it working again. Anyone else have the same problem???
type: bug,platform: Linux,status: needs triage
low
Major
2,796,586,042
PowerToys
Technical glitches while running the powertoy
### Microsoft PowerToys version 0.87.1.0 ### Installation method Microsoft Store ### Running as admin Yes ### Area(s) with issue? PowerToys Run, General ### Steps to reproduce when my pc is in idle state and after some time when i try to run the powertoys run using alt + space it don't register the action and fail to open ### ✔️ Expected Behavior when i open my pc with in 10 sec maximum it should open the powertoys run ### ❌ Actual Behavior It does not open the powertoys run and it some time it open directly using space button , also experienced at the same time that the it is unable to register my keyboard input and no keyword input seen sorry i don't have the screenshots right now ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,796,598,858
rust
Tracking issue for release notes of #133700: const-eval: detect more pointers as definitely not-null
This issue tracks the release notes text for #133700. ### Steps - [ ] Proposed text is drafted by PR author (or team) making the noteworthy change. - [ ] Issue is nominated for release team review of clarity for wider audience. - [ ] Release team includes text in release notes/blog posts. ### Release notes text The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing). ````markdown # Language - [const-eval: detect more pointers as definitely not-null based on their alignment](https://github.com/rust-lang/rust/pull/133700) ```` > [!TIP] > Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use. > The category will be de-duplicated with all the other ones by the release team. > > *More than one section can be included if needed.* ### Release blog section If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section. *Otherwise leave it empty.* ````markdown ```` cc @RalfJung, @lcnr -- origin issue/PR authors and assignees for starting to draft text
T-lang,T-compiler,relnotes,needs-triage,relnotes-tracking-issue
low
Minor
2,796,602,141
neovim
AI/LLM/GenAI related features
This is a stub issue, will update it over time. ## Problem Although "AI features" are not something we see a need for in Nvim core, I'm looking at the space and have noticed some "primitives" that are missing and require every plugin to reimplement things that are generally useful, even for non-AI plugins. ## Expected behavior - A way to mark a buffer or window as "busy" or "in progress". - Default 'statusline' shows the busy indicator (optionally progress "spinner"?). - https://github.com/neovim/neovim/issues/28809 - `vim.ui` interface for showing progress. - ✅ blink/highlight of a region will be helpful https://github.com/neovim/neovim/pull/32012 cc @siddhantdev - multiline completion https://github.com/neovim/neovim/issues/7769
ui,lua
low
Major
2,796,603,367
godot
[godot3.6] Compile web platform runtime error
### Tested versions 3.6 ### System information godot3.6 v3.6.1.rc.mono.custom_build [4c1514c7f] ### Issue description compile for the Web platform `scons platform=javascript tools=no target=release scons platform=javascript tools=no target=release_debug ` set compiled webassembly_debug.zip and webassembly_release.zip ![Image](https://github.com/user-attachments/assets/e709f2e0-c805-4191-b187-5538c20777a1) run web got a error modules\mono\glue\gd_glue.cpp:249 - Failed to export project: Missing BCL (Base Class Library) for platform: javascript ![Image](https://github.com/user-attachments/assets/660aed03-7105-4de6-a43f-29641158737e) ![Image](https://github.com/user-attachments/assets/5aee6efc-7566-4769-86ab-d64c4c17b2f2) ### Steps to reproduce - ### Minimal reproduction project (MRP) -
bug,platform:web,needs testing,topic:dotnet
low
Critical
2,796,610,756
PowerToys
Not able to copy certain info
### Microsoft PowerToys version 0.87.1 ### Installation method PowerToys auto-update ### Running as admin Yes ### Area(s) with issue? TextExtractor ### Steps to reproduce Tried to copy info. Text Extractor failed to copy. ![Image](https://github.com/user-attachments/assets/82602d10-57fa-4d60-8b57-b27306b2342b) ### ✔️ Expected Behavior Successful copy ### ❌ Actual Behavior Failed to copy particular info. Copied everything else on the page successfully. ### Other Software COMODO - Internet Security - 12.2.2.7036
Issue-Bug,Needs-Triage,Product-Text Extractor
low
Critical
2,796,611,015
react
[DevTools Bug] Cannot add child "301" to parent "155" because parent node was not found in the Store.
### Website or app www.github.com ### Repro steps The error was thrown at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1173126 at v.emit (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1140783) at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1142390 at bridgeListener (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1552662) ### How often does this bug happen? Every time ### DevTools package (automated) react-devtools-extensions ### DevTools version (automated) 6.0.1-c7c68ef842 ### Error message (automated) Cannot add child "301" to parent "155" because parent node was not found in the Store. ### Error call stack (automated) ```text at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1173126 at v.emit (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1140783) at chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1142390 at bridgeListener (chrome-extension://gpphkfbcpidddadnkolkpfckpihlkkil/build/main.js:1:1552662) ``` ### Error component stack (automated) ```text ``` ### GitHub query string (automated) ```text https://api.github.com/search/issues?q=Cannot add child to parent because parent node was not found in the Store. in:title is:issue is:open is:public label:"Component: Developer Tools" repo:facebook/react ```
Type: Bug,Status: Unconfirmed,Component: Developer Tools
medium
Critical
2,796,615,965
godot
[v4.4-dev7_win64] Failure to launch
### Tested versions Reproducible in: Godot_v4.4-dev7_win64, Godot_v4.4-beta1_win64 Not reproducible in: Godot_v4.4-dev6_win64 ### System information Windows 11 - RTX 4090 (Thunderbolt 4) - Driver 566.36 ### Issue description Upon running the Godot application, the cursor spins for a second, then stops without ever launching the application. When utilizing onboard graphics (Radeon 680M), this problem is not present. However, when utilizing an RTX 4090 over Thunderbolt 4 (eGPU), it occurs. Just to clarify, this is a full-size card, not a 4090 mobile. Logging doesn't appear out of the ordinary to me. If there is additional logging or information I can gather to assist, please do share. ``` PS C:\Users\kaboo\Desktop> .\Godot_v4.4-dev7_win64_console.exe --verbose WorkerThreadPool: 16 threads, 4 max low-priority. Godot Engine v4.4.dev7.official.46c8f8c5c - https://godotengine.org TextServer: Added interface "Dummy" TextServer: Added interface "ICU / HarfBuzz / Graphite (Built-in)" Native OpenGL API detected: 3.3: NVIDIA - NVIDIA GeForce RTX 4090 NVAPI: Init OK! NVAPI: Disabled OpenGL threaded optimization successfully NVAPI: Disabled G-SYNC for windowed mode successfully PS C:\Users\kaboo\Desktop> ``` ### Steps to reproduce Appears to possibly be hardware related. Steps are as simple as launching the application. ### Minimal reproduction project (MRP) Not Applicable
bug,platform:windows,crash,regression
low
Critical
2,796,626,997
rust
regression: ICE: Failed to normalize Alias(Opaque, AliasTy
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/Menestis.Skynet/log.txt ``` [INFO] [stdout] error: internal compiler error: compiler/rustc_middle/src/ty/normalize_erasing_regions.rs:168:90: Failed to normalize Alias(Opaque, AliasTy { args: [], def_id: DefId(0:850 ~ skynet[716d]::web::create_task::{opaque#0}::{opaque#0}), .. }), maybe try to call `try_normalize_erasing_regions` instead ```
I-ICE,P-low,T-compiler,regression-from-stable-to-beta
low
Critical
2,796,627,464
rust
regression: a value of type `HashMap<Pulse, u64>` cannot be built from
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/dimo414.advent-2023/log.txt ``` [INFO] [stderr] Compiling advent-2023 v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] error[E0277]: a value of type `HashMap<Pulse, u64>` cannot be built from an iterator over elements of type `&(Pulse, {integer})` [INFO] [stdout] --> src/bin/20/main.rs:278:88 [INFO] [stdout] | [INFO] [stdout] 278 | assert_eq!(config.counts, [(Pulse::Low, low), (Pulse::High, high)].into_iter().collect()); [INFO] [stdout] | ^^^^^^^ value of type `HashMap<Pulse, u64>` cannot be built from `std::iter::Iterator<Item=&(Pulse, {integer})>` [INFO] [stdout] 279 | } } ```
T-lang,T-compiler,regression-from-stable-to-beta,I-prioritize,I-lang-nominated,A-edition-2024
low
Critical
2,796,627,814
rust
regression: cannot borrow ... as immutable because it is also borrowed as mutable
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/LP-Finance-Inc.twamm/log.txt - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/Mokosha.pbrt_rust/log.txt - https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/gh/jw1912.diffable/log.txt ``` [INFO] [stdout] error[E0502]: cannot borrow `*inputs` as immutable because it is also borrowed as mutable [INFO] [stdout] --> examples/basic.rs:267:27 [INFO] [stdout] | [INFO] [stdout] 263 | if let Some(grd) = inputs[0].grad.as_mut() { [INFO] [stdout] | -------------- mutable borrow occurs here [INFO] [stdout] 264 | / *grd += output_grad [INFO] [stdout] 265 | | * if inputs[0].val > 0.0 { [INFO] [stdout] 266 | | 1.0 [INFO] [stdout] 267 | | } else if inputs[0].val == 0.0 { [INFO] [stdout] | | ^^^^^^^^^ immutable borrow occurs here [INFO] [stdout] ... | [INFO] [stdout] 270 | | -1.0 [INFO] [stdout] 271 | | }; [INFO] [stdout] | |_________________- mutable borrow later used here ```
T-compiler,regression-from-stable-to-beta,P-critical,T-types
low
Critical
2,796,627,979
rust
regression: ICE: `trimmed_def_paths` called, diagnostics were expected but none were emitted
- https://crater-reports.s3.amazonaws.com/beta-1.85-1/beta-2025-01-12/reg/quinn-proto-0.11.9/log.txt ``` [INFO] [stderr] `trimmed_def_paths` called, diagnostics were expected but none were emitted. Use `with_no_trimmed_paths` for debugging. This happened in the following `must_produce_diag` call's backtrace: [INFO] [stderr] 0: <rustc_errors::DiagCtxtHandle>::set_must_produce_diag [INFO] [stderr] 1: <rustc_session::session::Session>::record_trimmed_def_paths [INFO] [stderr] 2: rustc_middle::ty::print::pretty::trimmed_def_paths [INFO] [stderr] 3: rustc_query_impl::plumbing::__rust_begin_short_backtrace::<rustc_query_impl::query_impl::trimmed_def_paths::dynamic_query::{closure#2}::{closure#0}, rustc_middle::query::erase::Erased<[u8; 8]>> [INFO] [stderr] 4: rustc_query_system::query::plumbing::try_execute_query::<rustc_query_impl::DynamicConfig<rustc_query_system::query::caches::SingleCache<rustc_middle::query::erase::Erased<[u8; 8]>>, false, false, false>, rustc_query_impl::plumbing::QueryCtxt, false> [INFO] [stderr] 5: rustc_query_impl::query_impl::trimmed_def_paths::get_query_non_incr::__rust_end_short_backtrace ```
I-ICE,P-high,T-compiler,regression-from-stable-to-beta
low
Critical
2,796,677,342
flutter
I get this error in IOS : [ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PlatformException(channel-error, Unable to establish connection on channel: "dev.flutter.pigeon.path_provider_foundation.PathProviderApi.getDirectoryPath"., null, null) #0
When I try to run this in IOS I get this error. [ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PlatformException(channel-error, Unable to establish connection on channel: "dev.flutter.pigeon.path_provider_foundation.PathProviderApi.getDirectoryPath"., null, null) #0 PathProviderApi.getDirectoryPath (package:path_provider_foundation/messages.g.dart:95:7)
waiting for customer response,in triage
low
Critical
2,796,689,807
deno
Relax restrictions on scrypt `log_n`
Version: Deno 2.1.6 ```js // scrypt.js import crypto from 'node:crypto' crypto.scrypt('password', 'salt', 128, (err, hashed) => { if (err) throw err console.log('hashed = ', hashed.toString('hex')) }) crypto.scrypt('password', 'salt', 128, (err, hashed) => { if (err) throw err console.log('hashed = ', hashed.toString('hex')) }) ``` ```shell $ deno ./scrypt.js ============================================================ Deno has panicked. This is a bug in Deno. Please report this at https://github.com/denoland/deno/issues/new. If you can reliably reproduce this panic, include the reproduction steps and re-run with the RUST_BACKTRACE=1 env var set and include the backtrace in your report. Platform: linux x86_64 Version: 2.1.6 Args: ["/home/cyan/.asdf/installs/deno/2.1.4/bin/deno", "./scrypt.js"] thread 'tokio-runtime-worker' panicked at ext/node/ops/crypto/mod.rs:585:4: called `Result::unwrap()` on an `Err` value: InvalidParams note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace ============================================================ Deno has panicked. This is a bug in Deno. Please report this at https://github.com/denoland/deno/issues/new. If you can reliably reproduce this panic, include the reproduction steps and re-run with the RUST_BACKTRACE=1 env var set and include the backtrace in your report. Platform: linux x86_64 Version: 2.1.6 Args: ["/home/cyan/.asdf/installs/deno/2.1.4/bin/deno", "./scrypt.js"] thread 'tokio-runtime-worker' panicked at ext/node/ops/crypto/mod.rs:585:4: called `Result::unwrap()` on an `Err` value: InvalidParams ``` ```shell $ node ./scrypt.js hashed = 745731af4484f323968969eda289aeee005b5903ac561e64a5aca121797bf7734ef9fd58422e2e22183bcacba9ec87ba0c83b7a2e788f03ce0da06463433cda64176095fbbad7dc98c33fd75955b4b29c94f6e97617bd68d8ff17cf1ed5ad12f3fc6c8eb5b844f2d003ebaf5eaed19e8f665928472a6941f7efc6ebcdd6fd13a hashed = 745731af4484f323968969eda289aeee005b5903ac561e64a5aca121797bf7734ef9fd58422e2e22183bcacba9ec87ba0c83b7a2e788f03ce0da06463433cda64176095fbbad7dc98c33fd75955b4b29c94f6e97617bd68d8ff17cf1ed5ad12f3fc6c8eb5b844f2d003ebaf5eaed19e8f665928472a6941f7efc6ebcdd6fd13a ``` Here: https://github.com/denoland/deno/blob/57dd66ec3dae2a0fe376f6a43c476dfade421c84/ext/node/ops/crypto/mod.rs#L579-L585 The Result is unwrapped, and it will be an `Err` when len > 64 According to the Document here: https://docs.rs/scrypt/latest/scrypt/struct.Params.html#conditions But the same thing works in Node.js, maybe deno need to consider some alternatives?
upstream,node compat,crypto
low
Critical
2,796,720,352
godot
Duplicating resources and editing them without saving stores incorrect data in them and inspector displays wrong data.
### Tested versions Reproducible in: - v4.4.beta1.official [d33da79d3] - v4.3.stable.official [77dcf97d8] - v4.2.2.stable.official [15073afe3] ### System information Godot v4.3.stable - Windows 10.0.19045 - GLES3 (Compatibility) - NVIDIA GeForce GTX 750 Ti (NVIDIA; 32.0.15.6094) - Intel(R) Core(TM) i5-6400 CPU @ 2.70GHz (4 Threads) ### Issue description When setting up a resource, then duplicating that resource an setting it up (a few times) the resource file will have icorrect data in it but display the correct data in the editor. When the project is executed, the incorrect information will be used. https://github.com/user-attachments/assets/4ab695f9-bf9a-4c32-a93a-72d2f485346a ### Steps to reproduce 1. Create a Shortcut resource. 2. Assign it an InputKeyEvent 3. Duplicate that resource 4. Edit that resource InputKeyEvent to be something new. 5. (..repeat 2 or 3 times) 6. Run project. Shortcuts are assigned wrongly **Note:** When creating the shortcuts directly (Create New -> Resource) this didn't happen. Issue only occours when duplicating (Ctrl+D). The Minimal Reproduction Project can't be uploaded in its entirety since the editor will display the correctly stored data on reload. ### Minimal reproduction project (MRP) [shortcut_saving.zip](https://github.com/user-attachments/files/18462889/shortcut_saving.zip)
topic:editor,needs testing
low
Minor
2,796,746,011
tensorflow
TFnode on TensorflowonSpark 2.2.5
I'm running some code on Microsoft Fabric, and I use the following line: def map_fun(tf_args, ctx): cluster, server = TFNode.start_cluster_server(ctx) print('ctx') if ctx.job_name == "ps": server.join() else: print("Hello from worker", ctx.task_index) I'm getting an error regarding TFnode, does anyone know why this could be? TensorFlow version:2.12 TensorflowonSpark:2.2.5 The idea is to print a hello world to see if they are running on differents clusters.
type:support,TF 2.12
low
Critical
2,796,791,671
ollama
Embedding Model: iamgroot42/rover_nexus
Its the top model in MTEB Leaderboard.
model request
low
Minor
2,796,826,624
terminal
First turn off the performance option Show window contents when dragging. Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position.
### Windows Terminal version 1.21.3231.0 ### Windows build number 10.0.26100.2894 ### Other Software _No response_ ### Steps to reproduce First turn off the performance option Show window contents when dragging. Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position. ### Expected Behavior No black line appears when the window is zoomed out. ### Actual Behavior Open a Windows terminal window and drag the right edge of the window to zoom out the window. Black lines are displayed in the original position.
Issue-Bug,Needs-Author-Feedback,Needs-Triage,No-Recent-Activity
low
Major
2,796,831,254
godot
play_backwards Cannot Play Loop Animations
### Tested versions v4.4.beta1.official [4ce466d7f] ### System information Windows 11 (build 22631) - Multi-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2050 (NVIDIA; 32.0.15.6109) - 12th Gen Intel(R) Core(TM) i5-12500H (16 threads) ### Issue description I have encountered an issue where the `play_backwards()` method cannot play loop animations for models imported from Blender. When attempting to play these animations in reverse, they get stuck on the first frame, while all non-looping animations play correctly in reverse. ### Steps to reproduce Play a looped animation backwards ### Minimal reproduction project (MRP) N/A
bug,needs testing,topic:animation
low
Minor
2,796,841,741
terminal
Unable to Launch Windows Terminal via `wt.exe` or `windowsterminal.exe` After KB5050021 Update
### Windows Terminal version 1.21.3231.0 ### Windows build number 22621.4751 ### Other Software _No response_ ### Steps to reproduce ## Detailed Description After applying the KB5050021 update on January 14, 2025, the Windows Terminal fails to launch when executed via: - `wt.exe` in PowerShell, CMD, or VS Code integrated terminal. - Direct execution of `windowsterminal.exe` in `C:\Program Files\WindowsApps\Microsoft.WindowsTerminal_1.21.3231.0_x64__8wekyb3d8bbwe`. No error messages are displayed, and the process is not visible in `Task Manager` for long. The `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds. However, the terminal opens correctly when launched via its app icon in the Start menu. --- ## Steps to Reproduce 1. Update Windows 11 to build 22621.4751 via KB5050021. 2. Attempt to open Windows Terminal via `wt.exe` in a PowerShell or CMD window. 3. Alternatively, attempt to open `windowsterminal.exe` directly. 4. Observe that no terminal appears and no error messages are shown. --- ## Observations - When launched via the app icon, the terminal works correctly. - Running `Get-Process -Name *WindowsTerminal*, wt` after opening the terminal via the app icon shows `WindowsTerminal` as active. Running the same command after executing `wt.exe` returns no processes. - When executing `wt.exe`, the `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds. - The issue persists across the following environments: - Default PowerShell terminal. - CMD. - VS Code integrated terminal. - Resetting, uninstalling, and reinstalling Windows Terminal, as well as running `sfc /scannow` and DISM commands, did not resolve the issue. --- ## Expected Behavior `wt.exe` and `windowsterminal.exe` should launch Windows Terminal normally, as before the update. ## Additional Notes The issue affects workflows that depend on scripting or terminal automation via `wt.exe`. This problem has been consistently reproducible and may affect other users who installed the KB5050021 update. ### Expected Behavior ## Expected Behavior `wt.exe` and `windowsterminal.exe` should launch Windows Terminal normally, as before the update. --- ### Actual Behavior No error messages are displayed, and the process is not visible in `Task Manager` for long. The `WindowsTerminal.exe` process briefly appears in Task Manager but disappears after a couple of seconds. However, the terminal opens correctly when launched via its app icon in the Start menu.
Issue-Bug,Needs-Triage,Needs-Attention
low
Critical
2,796,865,708
storybook
[Bug]: create Nextjs sandbox failed.
### Describe the bug ``` ⚙️ Initializing Storybook > node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug ╭──────────────────────────────────────────────────────────────╮ │ │ │ Adding Storybook version 8.6.0-alpha.0 to your project.. │ │ This is a pre-release version. │ │ │ ╰──────────────────────────────────────────────────────────────╯ • Detecting project type. ✓ Installing dependencies... ➤ YN0000: · Yarn 4.6.0 ➤ YN0000: ┌ Resolution step ➤ YN0000: └ Completed ➤ YN0000: ┌ Fetch step ➤ YN0000: └ Completed ➤ YN0000: ┌ Link step ➤ YN0000: └ Completed ➤ YN0000: · Done in 0s 269ms • Adding Storybook support to your "Next" app ✔ Getting the correct version of 9 packages Configuring eslint-plugin-storybook in your package.json ⠧ Installing Storybook dependencies An error occurred while installing dependencies: YARN2 error YN0060: INCOMPATIBLE_PEER_DEPENDENCY -> @storybook/test is listed by your project with version 8.6.0-alpha.0 (p8d038), which doesn't satisfy what @storybook/react requests (but they have non-overlapping ranges!). YN0060: INCOMPATIBLE_PEER_DEPENDENCY -> storybook is listed by your project with version 8.6.0-alpha.0 (p23005), which doesn't satisfy what @storybook/addon-essentials and other dependencies request (but they have non-overlapping ranges!). YN0002: MISSING_PEER_DEPENDENCY -> before-storybook@workspace:. doesn't provide webpack (pd8fe4), requested by @storybook/nextjs. YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA -> Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code. YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA -> Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details. YN0001: EXCEPTION -> Error: Assertion failed: Writing attempt prevented to /Users/zhyd1997/workspace/storybook/code/addons/essentials/node_modules/@storybook/addon-docs which is outside project root: /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts Please check the logfile generated at ./storybook.log for troubleshooting and try again. Error: Error: YARN2 error YN0060: INCOMPATIBLE_PEER_DEPENDENCY -> @storybook/test is listed by your project with version 8.6.0-alpha.0 (p8d038), which doesn't satisfy what @storybook/react requests (but they have non-overlapping ranges!). YN0060: INCOMPATIBLE_PEER_DEPENDENCY -> storybook is listed by your project with version 8.6.0-alpha.0 (p23005), which doesn't satisfy what @storybook/addon-essentials and other dependencies request (but they have non-overlapping ranges!). YN0002: MISSING_PEER_DEPENDENCY -> before-storybook@workspace:. doesn't provide webpack (pd8fe4), requested by @storybook/nextjs. YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA -> Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code. YN0086: EXPLAIN_PEER_DEPENDENCIES_CTA -> Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details. YN0001: EXCEPTION -> Error: Assertion failed: Writing attempt prevented to /Users/zhyd1997/workspace/storybook/code/addons/essentials/node_modules/@storybook/addon-docs which is outside project root: /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts Please check the logfile generated at ./storybook.log for troubleshooting and try again. at Yarn2Proxy.addDependencies (/Users/zhyd1997/workspace/storybook/code/core/dist/common/index.cjs:25592:15) at async baseGenerator (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2166:5) at async generator4 (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2350:3) at async installStorybook (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:2966:12) at async doInitiate (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:3081:25) at async withTelemetry (/Users/zhyd1997/workspace/storybook/code/core/dist/core-server/index.cjs:51862:12) at async initiate (/Users/zhyd1997/workspace/storybook/code/lib/create-storybook/dist/bin/index.cjs:3146:26) An error occurred while executing: `node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug` 🚨 Initializing Storybook failed Error running task sandbox for nextjs/default-ts: { "shortMessage": "Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug", "command": "node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug", "escapedCommand": "\"node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug\"", "exitCode": 1, "failed": true, "timedOut": false, "isCanceled": false, "killed": false } Error: Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug at makeError (/Users/zhyd1997/workspace/storybook/scripts/node_modules/execa/lib/error.js:59:13) at handlePromise (/Users/zhyd1997/workspace/storybook/scripts/node_modules/execa/index.js:119:50) at process.processTicksAndRejections (node:internal/process/task_queues:95:5) at async exec (/Users/zhyd1997/workspace/storybook/scripts/utils/exec.ts:43:7) at async executeCLIStep (/Users/zhyd1997/workspace/storybook/scripts/utils/cli-step.ts:105:3) at async init (/Users/zhyd1997/workspace/storybook/scripts/tasks/sandbox-parts.ts:145:3) at async Object.run (/Users/zhyd1997/workspace/storybook/scripts/tasks/sandbox.ts:111:5) at async runTask (/Users/zhyd1997/workspace/storybook/scripts/task.ts:319:24) at async run (/Users/zhyd1997/workspace/storybook/scripts/task.ts:489:28) { shortMessage: 'Command failed with exit code 1: node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug', command: 'node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug', escapedCommand: '"node /Users/zhyd1997/workspace/storybook/code/lib/create-storybook/bin/index.cjs init --yes --debug"', exitCode: 1, signal: undefined, signalDescription: undefined, stdout: undefined, stderr: undefined, failed: true, timedOut: false, isCanceled: false, killed: false } ➜ storybook git:(next) ✗ ``` ### Reproduction link NA ### Reproduction steps 1. yarn task 2. select to generate `nextjs-default-ts` sandbox from template. 3. see the errors above. ### System ```bash Storybook Environment Info: System: OS: macOS 15.1.1 CPU: (14) arm64 Apple M3 Max Shell: 5.9 - /bin/zsh Binaries: Node: 18.16.0 - ~/.nvm/versions/node/v18.16.0/bin/node Yarn: 4.3.0 - ~/.nvm/versions/node/v18.16.0/bin/yarn <----- active npm: 9.5.1 - ~/.nvm/versions/node/v18.16.0/bin/npm Browsers: Chrome: 131.0.6778.266 Safari: 18.1.1 ``` ### Additional context ``` ➜ storybook git:(next) ✗ yarn task Installing Script Dependencies... ✔ Which task would you like to run? › Create the sandbox from a template (sandbox) ✔ What template would you like to make a sandbox for? › nextjs/default-ts To run this directly next time, use: yarn task --task sandbox --template nextjs/default-ts 👉 Selected sandbox: nextjs/default-ts Task readiness up to sandbox 🟢 install > 🟢 compile > 🟡 sandbox ✔ We need to run all tasks from sandbox onwards, would you like to start from an earlier task? › Create the sandbox from a template (sandbox) 🟢 install > 🟢 compile > 🔄 sandbox 👷 Bootstrapping Template > node /Users/zhyd1997/workspace/storybook/code/lib/cli-storybook/bin/index.cjs repro nextjs/default-ts --output /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts --no-init --debug ╭──────────────────────────────────────────────────╮ │ │ │ Creating a Storybook 8.6.0-alpha.0 sandbox.. │ │ This is a pre-release version. │ │ │ ╰──────────────────────────────────────────────────╯ ⚠️ /Users/zhyd1997/workspace/storybook/sandbox/nextjs-default-ts already exists! Overwriting... 🏃 Adding Next.js Latest (Webpack | ``` [storybook.log](https://github.com/user-attachments/files/18463756/storybook.log)
bug,good first issue,help wanted,cli,nextjs
low
Critical
2,796,865,913
puppeteer
[Feature]: Keyboard and Pointer Lock API Permission type added
### Feature description Can we please get the new chrome v131 Keyboard and Pointer Lock permissions added? https://pptr.dev/api/puppeteer.permission Current workaround requires using v23.7.1 for chrome v130 browser. https://pptr.dev/supported-browsers https://developer.chrome.com/blog/keyboard-lock-pointer-lock-permission We are currently unable to test our application on any version of puppeteer above v23.7.1 KeyboardLock and PointerLock were added to devtools-protocol in this commit (r1388031) https://github.com/ChromeDevTools/devtools-protocol/commit/e3771a854976cb9579b78c03c58bef4d5c63ca93
feature,P2,chrome,m133
low
Minor
2,796,877,194
tensorflow
Segmentation fault (core dumped) in `RaggedTensorToTensor`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.17 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tf.raw_ops.RaggedTensorToTensor` triggers crash. ### Standalone code to reproduce the issue ```shell import tensorflow as tf shape = tf.constant(-1, shape=[], dtype=tf.int64) values = tf.constant(0, shape=[0], dtype=tf.int32) default_value = tf.constant(0, shape=[], dtype=tf.int32) row_partition_tensors = tf.constant([0, 1, 6], shape=[3], dtype=tf.int64) row_partition_types = ["ROW_SPLITS"] tf.raw_ops.RaggedTensorToTensor( shape=shape, values=values, default_value=default_value, row_partition_tensors=[row_partition_tensors], row_partition_types=row_partition_types) ``` ### Relevant log output ```shell Segmentation fault (core dumped) ```
type:bug,comp:ops,2.17
medium
Critical
2,796,877,858
ui
[feat]: I would like to request a change to the cookie name used in the sidebar component.
### Feature description The current cookie name for the sidebar is as follows: ``` const SIDEBAR_COOKIE_NAME = "sidebar:state"; ``` However, this cookie name causes a validation error in msw, making it unusable. ``` TypeError: argument name is invalid ``` In msw, cookie names are validated against the following regular expression: msw.js ``` var cookieNameRegExp = /^[!#$%&'*+\-.^_`|~0-9A-Za-z]+$/; if (!cookieNameRegExp.test(name)) { throw new TypeError("argument name is invalid"); } ``` The issue was resolved when I changed the cookie name to "sidebar-state". Could you please adjust the cookie name to comply with this validation? Thank you! ### Affected component/components Sidebar ### Additional Context Additional details here... ### Before submitting - [x] I've made research efforts and searched the documentation - [x] I've searched for existing issues and PRs
area: request
low
Critical
2,796,881,366
vscode
GPU usage fluctuate in VS Code, but stably high in other IDE.
Type: <b>Performance Issue</b> I am building a BERT-like model with pytorch, but I encounter a very unstable GPU usage while training my model on VS Code. I tested the same code and python env in Spyder IDE, everything is fine and GPU usage is stably above 90%. So, it seems that there is something bog down my device and cause the bottleneck. VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Linux x64 6.8.0-51-generic Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Xeon(R) CPU E7-8890 v3 @ 2.50GHz (144 x 0)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off<br>webnn: disabled_off| |Load (avg)|0, 1, 12| |Memory (System)|125.72GB (118.82GB free)| |Process Argv|--crash-reporter-id 4e28e8db-6680-4dea-a577-845468db680c| |Screen Reader|no| |VM|0%| |DESKTOP_SESSION|ubuntu| |XDG_CURRENT_DESKTOP|Unity| |XDG_SESSION_DESKTOP|ubuntu| |XDG_SESSION_TYPE|x11| </details><details> <summary>Process Info</summary> ``` CPU % Mem MB PID Process 0 129 167128 code main 0 0 167132 zygote 0 129 167175 gpu-process 0 0 167210 broker 0 0 167133 zygote 0 0 167135 zygote 0 257 167191 window [1] (train.py - ECG_Encoder_for_JoLT - Visual Studio Code) 0 0 167177 utility-network-service 0 0 167318 ptyHost 0 0 167434 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh 0 0 167437 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh 0 0 167914 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh 0 0 168130 /usr/bin/bash --init-file /usr/share/code/resources/app/out/vs/workbench/contrib/terminal/common/scripts/shellIntegration-bash.sh 0 386 167333 extensionHost [1] 0 0 167632 /home/revlis_ai/.vscode/extensions/ms-python.python-2024.22.2-linux-x64/python-env-tools/bin/pet server 0 0 168094 /usr/share/code/code /usr/share/code/resources/app/extensions/json-language-features/server/dist/node/jsonServerMain --node-ipc --clientProcessId=167333 0 257 179521 electron-nodejs (bundle.js ) 0 129 167364 shared-process 0 0 180812 /bin/sh -c /usr/bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command= 0 0 180813 /usr/bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command= 0 129 167366 fileWatcher [1] ``` </details> <details> <summary>Workspace Info</summary> ``` | Window (train.py - ECG_Encoder_for_JoLT - Visual Studio Code) | Folder (ECG_Encoder_for_JoLT): more than 44316 files | File types: dat(9641) hea(9641) png(610) npz(19) hdf5(18) txt(8) pt(8) | py(8) pyc(3) csv(3) | Conf files:; ``` </details> <details><summary>Extensions (23)</summary> Extension|Author (truncated)|Version ---|---|--- arepl|alm|3.0.0 githistory|don|0.6.20 gitlens|eam|16.2.0 vscode-html-css|ecm|2.0.12 copilot|Git|1.257.0 copilot-chat|Git|0.23.2 vscode-docker|ms-|1.29.3 vscode-language-pack-zh-hant|MS-|1.96.2024121109 debugpy|ms-|2024.14.0 python|ms-|2024.22.2 vscode-pylance|ms-|2024.12.1 datawrangler|ms-|1.16.0 jupyter|ms-|2024.11.0 jupyter-keymap|ms-|1.1.2 jupyter-renderers|ms-|1.0.21 vscode-jupyter-cell-tags|ms-|0.1.9 vscode-jupyter-slideshow|ms-|0.1.6 remote-containers|ms-|0.394.0 remote-wsl|ms-|0.88.5 vscode-yaml|red|1.15.0 pdf|tom|1.2.2 vscode-pets|ton|1.27.0 tconda|zhp|0.0.5 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 vscod805cf:30301675 binariesv615:30325510 vsaa593cf:30376535 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 a9j8j154:30646983 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 h48ei257:31000450 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupyter:31046869 2f103344:31071589 nativerepl2:31139839 pythonrstrctxt:31112756 nativeloc1:31192215 cf971741:31144450 iacca1:31171482 notype1:31157159 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 6074i472:31201624 dwoutputs:31217127 ``` </details> <!-- generated by issue reporter -->
info-needed
low
Critical
2,796,881,534
tensorflow
Aborted in `tf.raw_ops.RaggedGather`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.17 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tf.raw_ops.RaggedGather` triggers crash. ### Standalone code to reproduce the issue ```shell params_nested_splits = tf.constant(0, shape=[3], dtype=tf.int64) params_dense_values = tf.constant(1, shape=[0], dtype=tf.float32) indices = tf.constant(0, shape=[], dtype=tf.int64) OUTPUT_RAGGED_RANK = 1 PARAMS_RAGGED_RANK = 1 tf.raw_ops.RaggedGather( params_nested_splits=[params_nested_splits], params_dense_values=params_dense_values, indices=indices, OUTPUT_RAGGED_RANK=1, name=None ) ``` ### Relevant log output ```shell 2025-01-18 09:30:00.549762: F tensorflow/core/framework/tensor.cc:844] Check failed: dtype() == expected_dtype (9 vs. 1) float expected, got int64 Aborted (core dumped) ```
type:bug,comp:ops,2.17
medium
Critical
2,796,891,774
flutter
'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 842 pos 12: '_hold == null': is not true.
### Steps to reproduce I wanted to create an instant return effect so that the header would always be open or closed. Essentially this code works, but at some point it produces a critical error and scrolling stops working. ### Expected results . ### Actual results / ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); @override Widget build(BuildContext context) { return MaterialApp( darkTheme: ThemeData.dark(), themeMode: ThemeMode.dark, ); } } class TelegramStyleAppBar extends StatefulWidget { const TelegramStyleAppBar({super.key}); @override _TelegramStyleAppBarState createState() => _TelegramStyleAppBarState(); } class _TelegramStyleAppBarState extends State<TelegramStyleAppBar> { late ScrollController _scrollController; double expandedHeight = 300.0; double collapsedHeight = 100.0; @override void initState() { super.initState(); _scrollController = ScrollController(); // Инициализация слушателей WidgetsBinding.instance.addPostFrameCallback( (timeStamp) { print(_scrollController.position); _scrollController.position.isScrollingNotifier.addListener( () { if (!_scrollController.position.isScrollingNotifier.value) { _handleSnapScroll(); } }, ); }, ); } @override void dispose() { _scrollController.dispose(); super.dispose(); } @override Widget build(BuildContext context) { return Scaffold( backgroundColor: Colors.black, body: CustomScrollView( controller: _scrollController, physics: const BouncingScrollPhysics(), slivers: [ SliverAppBar( pinned: true, stretch: true, surfaceTintColor: Colors.transparent, expandedHeight: expandedHeight, collapsedHeight: collapsedHeight, backgroundColor: Colors.black, flexibleSpace: LayoutBuilder( builder: (BuildContext context, BoxConstraints constraints) { final appBarHeight = constraints.biggest.height; final normalizedHeight = (appBarHeight - collapsedHeight) / (expandedHeight - collapsedHeight); final clampedHeight = normalizedHeight.clamp(0.0, 1.0); return FlexibleSpaceBar( centerTitle: true, title: Opacity( opacity: 1, child: Text( 'Telegram Header', style: TextStyle( fontSize: 20 + clampedHeight * 10, color: Colors.white, ), ), ), background: Container( color: Colors.blue, child: Center( child: Text( 'Pull to Stretch', style: TextStyle( fontSize: 16 + clampedHeight * 24, color: Colors.white, ), ), ), ), ); }, ), ), SliverList( delegate: SliverChildBuilderDelegate( (context, index) => ListTile( title: Text('Item $index'), ), childCount: 50, ), ), ], ), ); } void _handleSnapScroll() async { double offset = _scrollController.offset; double snapThreshold = (expandedHeight - collapsedHeight) / 2; // Проверка, на какой позиции находится скролл if (offset < snapThreshold) { // Скролл до раскрытой шапки await _scrollController.animateTo( 0.0, duration: const Duration(milliseconds: 300), curve: Curves.easeInOut, ); } else if (offset < expandedHeight - collapsedHeight) { // Скролл до свернутой шапки await _scrollController.animateTo( expandedHeight - collapsedHeight, duration: const Duration(milliseconds: 300), curve: Curves.easeInOut, ); } } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] </details> ### Logs <details open><summary>Logs</summary> ```console Restarted application in 432ms. flutter: ScrollPositionWithSingleContext#5d946(offset: 0.0, range: 0.0..2288.0, viewport: 874.0, ScrollableState, BouncingScrollPhysics -> BouncingScrollPhysics -> RangeMaintainingScrollPhysics, IdleScrollActivity#a82a1, ScrollDirection.idle) [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. [ERROR:flutter/impeller/base/validation.cc(59)] Break on 'ImpellerValidationBreak' to inspect point of failure: Could not find font in the atlas. ════════ Exception caught by gesture ═══════════════════════════════════════════ The following assertion was thrown while handling a gesture: 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 842 pos 12: '_hold == null': is not true. Either the assertion indicates an error in the framework itself, or we should provide substantially more information in this error message to help you determine and fix the underlying cause. In either case, please report this assertion by filing a bug on GitHub: https://github.com/flutter/flutter/issues/new?template=2_bug.yml When the exception was thrown, this was the stack: #2 ScrollableState._handleDragStart (package:flutter/src/widgets/scrollable.dart:842:12) #3 DragGestureRecognizer._checkStart.<anonymous closure> (package:flutter/src/gestures/monodrag.dart:790:53) #4 GestureRecognizer.invokeCallback (package:flutter/src/gestures/recognizer.dart:351:24) #5 DragGestureRecognizer._checkStart (package:flutter/src/gestures/monodrag.dart:790:7) #6 DragGestureRecognizer._checkDrag (package:flutter/src/gestures/monodrag.dart:757:5) #7 DragGestureRecognizer.acceptGesture (package:flutter/src/gestures/monodrag.dart:684:7) #8 GestureArenaManager._resolveByDefault (package:flutter/src/gestures/arena.dart:271:25) #9 GestureArenaManager._tryToResolveArena.<anonymous closure> (package:flutter/src/gestures/arena.dart:251:31) (elided 4 frames from class _AssertionError and dart:async) Handler: "onStart" Recognizer: VerticalDragGestureRecognizer#9732c start behavior: start ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 847 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by gesture ═══════════════════════════════════════════ 'package:flutter/src/widgets/scrollable.dart': Failed assertion: line 853 pos 12: '_hold == null || _drag == null': is not true. ════════════════════════════════════════════════════════════════════════════════ Reloaded 1 of 709 libraries in 394ms (compile: 38 ms, reload: 161 ms, reassemble: 135 ms). Reloaded 1 of 709 libraries in 328ms (compile: 43 ms, reload: 103 ms, reassemble: 91 ms). ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console f% a001@MacBook-exeshka telegram_ui % flutter doctor -v [✓] Flutter (Channel stable, 3.24.5, on macOS 15.2 24C101 darwin-arm64, locale ru-KZ) • Flutter version 3.24.5 on channel stable at /Users/a001/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision dec2ee5c1f (9 weeks ago), 2024-11-13 11:13:06 -0800 • Engine revision a18df97ca5 • Dart version 3.5.4 • DevTools version 2.37.3 [✓] Android toolchain - develop for Android devices (Android SDK version 35.0.0) • Android SDK at /Users/a001/Library/Android/sdk • Platform android-35, build-tools 35.0.0 • Java binary at: /opt/homebrew/Cellar/openjdk@17/17.0.13/libexec/openjdk.jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment Homebrew (build 17.0.13+0) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 16.2) • Xcode at /Applications/Xcode.app/Contents/Developer • Build 16C5032a • CocoaPods version 1.16.2 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 2024.2) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 21.0.3+-79915917-b509.11) [✓] VS Code (version 1.95.3) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.102.0 [✓] Connected device (5 available) • iPhone exeshka (mobile) • 00008101-0015148A0C40001E • ios • iOS 18.2 22C150 • iPhone 16 Pro (mobile) • 54FA6352-39FF-44D9-8BB9-ADF8BCF9E69E • ios • com.apple.CoreSimulator.SimRuntime.iOS-18-2 (simulator) • macOS (desktop) • macos • darwin-arm64 • macOS 15.2 24C101 darwin-arm64 • Mac Designed for iPad (desktop) • mac-designed-for-ipad • darwin • macOS 15.2 24C101 darwin-arm64 • Chrome (web) • chrome • web-javascript • Google Chrome 131.0.6778.266 [✓] Network resources • All expected network resources are available. • No issues found! a001@MacBook-exeshka telegram_ui % ``` </details>
waiting for customer response,in triage
low
Critical
2,796,900,433
vscode
Explorer not working as intended
Type: <b>Bug</b> The explorer doesnt show the folders and folder structure. VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Windows_NT x64 10.0.22631 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|11th Gen Intel(R) Core(TM) i5-11300H @ 3.10GHz (8 x 3110)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|undefined| |Memory (System)|7.75GB (0.39GB free)| |Process Argv|--crash-reporter-id f57f71a7-a455-4bb6-87be-c0f886d15903| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (4)</summary> Extension|Author (truncated)|Version ---|---|--- prettier-vscode|esb|11.0.0 copilot|Git|1.257.0 copilot-chat|Git|0.23.2 LiveServer|rit|5.7.9 </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vspor879:30202332 vspor708:30202333 vspor363:30204092 vscod805:30301674 binariesv615:30325510 vsaa593cf:30376535 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 2e7ec940:31000449 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupytercf:31046870 nativerepl1:31139838 pythonrstrctxt:31112756 nativeloc1:31192215 cf971741:31144450 iacca1:31171482 notype1cf:31157160 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 3d9ag387:31215808 6074i472:31201624 dwoutputs:31217127 ``` </details> <!-- generated by issue reporter -->
info-needed
low
Critical
2,796,906,847
pytorch
Bracket indexing not working
### 🐛 Describe the bug Unsqueezing not working ```python import torch tn = torch.randn(6980, 1, 12, 16, 20) tn[[1], :, :, :, :].shape # (1, 1, 12, 16, 20) tn[[1], :, [11], :, :].shape # (1, 1, 16, 20) ``` ### Versions PyTorch version: N/A Is debug build: N/A CUDA used to build PyTorch: N/A ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.4 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.12.4 | packaged by Anaconda, Inc. | (main, Jun 18 2024, 15:12:24) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-6.8.0-51-generic-x86_64-with-glibc2.35 Is CUDA available: N/A CUDA runtime version: 11.5.119 CUDA_MODULE_LOADING set to: N/A GPU models and configuration: GPU 0: NVIDIA GeForce RTX 4090 GPU 1: NVIDIA GeForce RTX 4090 Nvidia driver version: 555.42.02 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: N/A CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 39 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Vendor ID: GenuineIntel Model name: 13th Gen Intel(R) Core(TM) i9-13900F CPU family: 6 Model: 183 Thread(s) per core: 2 Core(s) per socket: 24 Socket(s): 1 Stepping: 1 CPU max MHz: 5600.0000 CPU min MHz: 800.0000 BogoMIPS: 3993.60 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb intel_pt sha_ni xsaveopt xsavec xgetbv1 xsaves split_lock_detect user_shstk avx_vnni dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp hwp_pkg_req hfi vnmi umip pku ospke waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm md_clear serialize arch_lbr ibt flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 896 KiB (24 instances) L1i cache: 1.3 MiB (24 instances) L2 cache: 32 MiB (12 instances) L3 cache: 36 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-31 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Mitigation; Clear Register File Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI BHI_DIS_S Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] numpy==2.1.0 [conda] blas 1.0 mkl [conda] cuda-cudart 12.1.105 0 nvidia [conda] cuda-cupti 12.1.105 0 nvidia [conda] cuda-libraries 12.1.0 0 nvidia [conda] cuda-nvrtc 12.1.105 0 nvidia [conda] cuda-nvtx 12.1.105 0 nvidia [conda] cuda-opencl 12.3.52 0 nvidia [conda] cuda-runtime 12.1.0 0 nvidia [conda] easy-torch 1.3.2 pypi_0 pypi [conda] ffmpeg 4.3 hf484d3e_0 pytorch [conda] libcublas 12.1.0.26 0 nvidia [conda] libcufft 11.0.2.4 0 nvidia [conda] libcurand 10.3.4.52 0 nvidia [conda] libcusolver 11.4.4.55 0 nvidia [conda] libcusparse 12.0.2.55 0 nvidia [conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch [conda] libnvjitlink 12.1.105 0 nvidia [conda] mkl 2023.1.0 h213fc3f_46343 [conda] mkl-service 2.4.0 py311h5eee18b_1 [conda] mkl_fft 1.3.8 py311h5eee18b_0 [conda] mkl_random 1.2.4 py311hdb19cb5_0 [conda] numpy 1.24.4 pypi_0 pypi [conda] nvidia-cublas-cu12 12.1.3.1 pypi_0 pypi [conda] nvidia-cuda-cupti-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cuda-nvrtc-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cuda-runtime-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cudnn-cu12 8.9.2.26 pypi_0 pypi [conda] nvidia-cufft-cu12 11.0.2.54 pypi_0 pypi [conda] nvidia-curand-cu12 10.3.2.106 pypi_0 pypi [conda] nvidia-cusolver-cu12 11.4.5.107 pypi_0 pypi [conda] nvidia-cusparse-cu12 12.1.0.106 pypi_0 pypi [conda] nvidia-nccl-cu12 2.20.5 pypi_0 pypi [conda] nvidia-nvjitlink-cu12 12.4.127 pypi_0 pypi [conda] nvidia-nvtx-cu12 12.1.105 pypi_0 pypi [conda] pytorch-cuda 12.1 ha16c6d3_5 pytorch [conda] pytorch-forecasting 1.2.0 pypi_0 pypi [conda] pytorch-lightning 2.2.0 pypi_0 pypi [conda] pytorch-mutex 1.0 cuda pytorch [conda] torch 2.3.0 pypi_0 pypi [conda] torch-cluster 1.6.3+pt23cu121 pypi_0 pypi [conda] torch-geometric 2.4.0 pypi_0 pypi [conda] torch-scatter 2.1.2+pt23cu121 pypi_0 pypi [conda] torch-sparse 0.6.18+pt23cu121 pypi_0 pypi [conda] torch-spline-conv 1.2.2+pt23cu121 pypi_0 pypi [conda] torch-summary 1.4.5 pypi_0 pypi [conda] torchaudio 2.3.0 pypi_0 pypi [conda] torchinfo 1.8.0 pypi_0 pypi [conda] torchmetrics 1.3.0.post0 pypi_0 pypi [conda] torchsummary 1.5.1 pypi_0 pypi [conda] torchvision 0.18.0 pypi_0 pypi [conda] triton 2.3.0 pypi_0 pypi
needs reproduction,triaged,module: advanced indexing
low
Critical
2,796,928,799
pytorch
Please add fp16 to MPS devices.
### 🚀 The feature, motivation and pitch I used torch==2.7 to train llama via huggingface transformers,but Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 9.00it/s] trainable params: 4,194,304 || all params: 6,933,450,752 || trainable%: 0.0605 Traceback (most recent call last): File "/Users/rbhan/Data/StellAIHub/Train/LanguageModel/xmt.py", line 115, in <module> trainer = Trainer( ^^^^^^^^ File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/utils/deprecation.py", line 165, in wrapped_func return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/trainer.py", line 459, in __init__ self.create_accelerator_and_postprocess() File "/Users/rbhan/Data/AIHub/Trans-Penv/transformers/src/transformers/trainer.py", line 5071, in create_accelerator_and_postprocess self.accelerator = Accelerator(**args) ^^^^^^^^^^^^^^^^^^^ File "/Users/rbhan/Data/AIHub/Trans-Penv/accelerate/src/accelerate/accelerator.py", line 495, in __init__ raise ValueError(f"fp16 mixed precision requires a GPU (not {self.device.type!r}).") ValueError: fp16 mixed precision requires a GPU (not 'mps'). ### Alternatives So when can pytorch support fp16 on MPS?I have waited for one year,but no solution.Thank you! ### Additional context _No response_ cc @mcarilli @ptrblck @leslie-fang-intel @jgong5 @kulinseth @albanD @malfet @DenisVieriu97 @jhavukainen
needs reproduction,triaged,module: amp (automated mixed precision),module: mps
low
Critical
2,796,933,056
tensorflow
Aborted in ` tf.nn.conv3d_transpose`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? No ### Source source ### TensorFlow version tf 2.16.1 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tf.nn.conv3d_transpose` triggers crash. ### Standalone code to reproduce the issue ```shell https://colab.research.google.com/drive/1wA5pUfm4HpNpHx-PawrZt4CKS2YCUR-4?usp=sharing ``` ### Relevant log output ```shell 2025-01-18 11:31:33.696896: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2754] Non-OK-status: GetNodeAttr(orig_node->def(), "strides", &strides) status: INVALID_ARGUMENT: Attr strides has value 5717580618211388939 out of range for an int32 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:apis,comp:ops,TF 2.16
medium
Critical
2,796,938,285
godot
project manager unable to resolve icon referenced by uid before import
### Tested versions 4.4-beta1 ### System information (Native) Godot v4.4.beta1 - Windows 10 (build 19045) - Multi-window, 2 monitors - OpenGL 3 (Compatibility) - NVIDIA GeForce RTX 4070 (NVIDIA; 32.0.15.6636) - AMD Ryzen 9 3950X 16-Core Processor (32 threads) (WSL2 on above system) Godot v4.4.beta1 - Ubuntu 24.04.1 LTS 24.04 - X11 display driver, Multi-window, 2 monitors - OpenGL 3 (Compatibility) - D3D12 (NVIDIA GeForce RTX 4070) - AMD Ryzen 9 3950X 16-Core Processor (32 threads) ### Issue description The project manager is unable to resolve an icon referenced by uid before first import: ``` ERROR: Unrecognized UID: "uid://b3xkos3u4jns4". at: get_id_path (core/io/resource_uid.cpp:157) ERROR: Error opening file ''. at: load_image (core/io/image_loader.cpp:89) ``` ### Steps to reproduce open project manager import MRP without "Edit Now" -> no icon, errors in console ### Minimal reproduction project (MRP) [icontest44b1.zip](https://github.com/user-attachments/files/18464271/icontest44b1.zip)
bug,topic:editor
low
Critical
2,796,939,108
godot
GPUParticles3D collide with their gizmo icon, which can make them appear to not move at all if using the Point emission mode
### Tested versions - Reproducible in: 4.4.beta1, 4.4.beta 7b1ed520b - Not reproducible in: 4.3.stable ### System information Godot v4.4.beta (7b1ed520b) - Fedora Linux 41 (KDE Plasma) on X11 - X11 display driver, Multi-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 4090 (nvidia; 565.77) - 13th Gen Intel(R) Core(TM) i9-13900K (32 threads) ### Issue description GPUParticles3D collide with their gizmo icon, which can make them appear to not move at all if using the Point emission mode. I bisected the regression to https://github.com/godotengine/godot/pull/91174. This likely occurs because making the gizmo material use alpha scissor means it's now taken into account by the heightfield collision generator. We can probably fix this by changing the cull mask of the GPUParticlesCollisionHeightField3D to exclude the editor-only gizmo layers (`Node3DEditorViewport::GIZMO_EDIT_LAYER`). This is further proven by the collision's shape matching the GPUParticles3D gizmo icon. You can see this by using a sphere emission shape and greatly increasing the particle count: ![Image](https://github.com/user-attachments/assets/c769a37c-a58b-408f-8c55-623bfe8e9c9e) Relevant code: https://github.com/godotengine/godot/blob/7b1ed520bdcca21d36e25e2094802b9b33dae2c6/editor/plugins/node_3d_editor_gizmos.cpp#L831-L837 https://github.com/godotengine/godot/blob/7b1ed520bdcca21d36e25e2094802b9b33dae2c6/scene/3d/gpu_particles_collision_3d.cpp ### Steps to reproduce - Set up a scene with various meshes and particles that have collision enabled. - Add a GPUParticlesCollisionHeightField3D node. - Save the scene and use **Scene > Reload Saved Scene**. - Notice how particles no longer move due to colliding with their gizmo. ### Minimal reproduction project (MRP) [test_particle_collision_gizmo.zip](https://github.com/user-attachments/files/18464280/test_particle_collision_gizmo.zip)
bug,topic:editor,regression,topic:3d,topic:particles
low
Minor
2,796,940,106
pytorch
[RFC] Improve performance for layer_norm op for cuda with revectorized
### 🚀 The feature, motivation and pitch I found there is big perf drop if the size of layer_norm's inside size is not multiple of 4, there is a micro test case: ```python import torch DEVICE=torch.device('cuda') # Time cost for near 1024 for cnt in range(2040, 2050): x = torch.randn(4096, cnt, device=DEVICE, dtype=torch.float32) w_shape = (x.shape[-1], ) #warm up need_warmup = True round = 5 if need_warmup: for _ in range(round): output = torch.nn.functional.layer_norm(x, w_shape) torch.cuda.synchronize() start_time = torch.cuda.Event(enable_timing=True) end_time = torch.cuda.Event(enable_timing=True) # Start time start_time.record() # Apply layernorm for _ in range(round): output = torch.nn.functional.layer_norm(x, w_shape) # End time end_time.record() torch.cuda.synchronize() # Calculate elapsed time elapsed_time_ms = start_time.elapsed_time(end_time) # print(f"CUDA Time: {elapsed_time_ms:.6f} ms") gbps = lambda ms: round * 2 * x.numel() * x.element_size() * 1e-9 / (ms * 1e-3) print(f"n as {cnt} of softmax: {gbps(elapsed_time_ms):.6f} gb/s") ``` Its output is: ``` n as 2040 of softmax: 483.555543 gb/s n as 2041 of softmax: 345.531858 gb/s n as 2042 of softmax: 345.369984 gb/s n as 2043 of softmax: 347.825623 gb/s n as 2044 of softmax: 489.580822 gb/s n as 2045 of softmax: 345.591973 gb/s n as 2046 of softmax: 346.057951 gb/s n as 2047 of softmax: 345.850055 gb/s n as 2048 of softmax: 470.192376 gb/s n as 2049 of softmax: 347.012446 gb/s ``` We could see for the perf for input with `N = 2040, 2044, 2048` is obvious greater than other input `(480 vs 340)`. Therefore, what I would like to do is **mitigating the perf gap between these input**. ### Alternatives The root cause is that there is two kernels for `layer_norm`: https://github.com/pytorch/pytorch/blob/5e4cf3e6ad6f1f06436f409b394ae02e5ed5583d/aten/src/ATen/native/cuda/layer_norm_kernel.cu#L824-L837 We could see if `N` is multiple of `num_vec_elems (4)`, it will call a kernel called `launch_vectorized_layer_norm_kernel`, which could load vectorized elements. So what we could do is to also enable vectorized elements load for those case whose `N` is not multiple for `num_vec_elems (4)`. I tried a draft implement to achieve it, it could improve performance same as the vectorized case, for this case it's about **~40%**, `(480 vs 340) take 2042 as example`. Optimized data: ``` n as 2040 of softmax: 459.758796 gb/s n as 2041 of softmax: 497.804886 gb/s n as 2042 of softmax: 479.061575 gb/s n as 2043 of softmax: 477.197096 gb/s n as 2044 of softmax: 473.285078 gb/s n as 2045 of softmax: 473.795206 gb/s n as 2046 of softmax: 495.999985 gb/s n as 2047 of softmax: 499.649134 gb/s n as 2048 of softmax: 455.111104 gb/s n as 2049 of softmax: 473.928423 gb/s ``` ### Additional context Here is a [doc](https://docs.google.com/document/d/1TFIbJAO3tek1-EltVvMC_0TgYuEeBZw8rxDrSb_Wx8Q/edit?tab=t.0) contains some stuffs about it. And there is another optimization, we could see there is still perf gap for `layer_norm` in `pytorch` and `triton`, we may mitigate the gap by using **register to cache the data in pytorch kernel**, because the `launch_vectorized_layer_norm_kernel` load data from gmem twice yet, but I guess the `N` for `layer_norm` op may usually be a very big number (> 10k for some 2d layer_norm), so it may introduce much register pressure, WDYT? cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @ptrblck @msaroufim @eqy
module: nn,module: cuda,triaged,topic: performance
low
Major
2,796,945,049
tensorflow
Abort in `tensorflow.keras.backend.conv2d`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? No ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.keras.backend.conv2d` triggers crash. The crash occurs in the operator Conv2D ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow x = np.array([1.5841993e+38, 4.7480855e+36, -1.8916006e+38], dtype=np.float32) kernel = np.ones((13, 8, 5, 14), dtype=np.float32) tensorflow.keras.backend.conv2d(x=x, kernel=kernel) ``` ### Relevant log output ```shell 2025-01-18 12:02:15.375499: F ./tensorflow/core/util/tensor_format.h:428] Check failed: index >= 0 && index < num_total_dims Invalid index from the dimension: 3, 0, C Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:apis,2.17
medium
Critical
2,796,947,730
tensorflow
Aborted (core dumped) in `tensorflow.nn.max_pool1d`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.nn.max_pool1d` triggers crash. The crash occurs in the operator MaxPool ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = np.ones((17,7,14), dtype=np.float32) ksize = [3134755050008138495] strides = [48732] padding = 'SAME' tensorflow.nn.max_pool1d(input=input, ksize=ksize, strides=strides, padding=padding) ``` ### Relevant log output ```shell 2025-01-18 12:09:37.053592: F tensorflow/core/common_runtime/mkl_layout_pass.cc:1609] Non-OK-status: GetNodeAttr(n->def(), "ksize", &ksize) status: INVALID_ARGUMENT: Attr ksize has value 3134755050008138495 out of range for an int32 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,949,006
tensorflow
Floating point exception in `tensorflow.nn.max_pool1d`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.nn.max_pool1d` triggers crash. The crash occurs in the operator MaxPool ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = np.ones((16,19,18), dtype=np.float32) ksize = [58486] strides = [49657] padding = 'VALID' tensorflow.nn.max_pool1d(input=input, ksize=ksize, strides=strides, padding=padding) ``` ### Relevant log output ```shell Floating point exception (core dumped) ```
stat:awaiting response,type:bug,comp:apis,comp:ops,2.17
medium
Critical
2,796,950,022
tensorflow
Aborted in `tensorflow.nn.max_pool3d`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.nn.max_pool3d` triggers crash. The crash occurs in the operator MaxPool3D ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = np.ones((20, 19, 20, 5, 0), dtype=np.float32) ksize = [3505357736] strides = [2230364274603370470] padding = 'SAME' tensorflow.nn.max_pool3d(input=input, ksize=ksize, strides=strides, padding=padding) ``` ### Relevant log output ```shell 2025-01-18 12:15:49.067137: F tensorflow/core/common_runtime/mkl_layout_pass.cc:1609] Non-OK-status: GetNodeAttr(n->def(), "ksize", &ksize) status: INVALID_ARGUMENT: Attr ksize has value 3505357736 out of range for an int32 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,951,107
rust
Tracking Issue for `stdarch_s390x`
Feature gate: `#![feature(stdarch_s390x)]` This is a tracking issue for the s390x (aka SystemZ) intrinsics in `core::arch::s390x`. ### Public API Everything in `core::arch::s390x`. Missing instructions based on the GCC [vecintrin.h](https://github.com/gcc-mirror/gcc/blob/master/gcc/config/s390/vecintrin.h) - [ ] `vec_step` - [ ] `__lcbb` - [ ] `vec_adde` - [ ] `vec_addec` - [ ] `vec_all_nle` - [ ] `vec_all_nlt` - [ ] `vec_any_nle` - [ ] `vec_any_nlt` - [ ] `vec_genmask` - [ ] `vec_genmasks_8` - [ ] `vec_genmasks_16` - [ ] `vec_genmasks_32` - [ ] `vec_genmasks_64` - [ ] `vec_splat_u8` - [ ] `vec_splat_s8` - [ ] `vec_splat_u16` - [ ] `vec_splat_s16` - [ ] `vec_splat_u32` - [ ] `vec_splat_s32` - [ ] `vec_splat_u64` - [ ] `vec_splat_s64` - [ ] `vec_sube` - [ ] `vec_subec` - [ ] `vec_checksum` - [ ] `vec_gfmsum_128` - [ ] `vec_gfmsum_accum_128` - [ ] `vec_ceil` - [ ] `vec_roundp` - [ ] `vec_floor` - [ ] `vec_roundm` - [ ] `vec_trunc` - [ ] `vec_roundz` - [ ] `vec_rint` - [ ] `vec_roundc` - [ ] `vec_round` - [ ] `vec_doublee` - [ ] `vec_floate` - [ ] `vec_all_nan` - [ ] `vec_all_numeric` - [ ] `vec_any_nan` - [ ] `vec_any_numeric` - [ ] `vec_extend_to_fp32_hi` - [ ] `vec_extend_to_fp32_lo` - [ ] `vec_round_from_fp32` - [ ] `vec_convert_to_fp16` - [ ] `vec_convert_from_fp16` - [ ] `vec_gen_element_masks_8` - [ ] `vec_gen_element_masks_16` - [ ] `vec_gen_element_masks_32` - [ ] `vec_gen_element_masks_64` - [ ] `vec_gen_element_masks_128` - [x] `vec_abs` - [ ] `vec_add_u128` - [ ] `vec_addc` - [ ] `vec_addc_u128` - [ ] `vec_adde_u128` - [ ] `vec_addec_u128` - [ ] `vec_all_eq` - [ ] `vec_all_ge` - [ ] `vec_all_gt` - [ ] `vec_all_le` - [ ] `vec_all_lt` - [ ] `vec_all_ne` - [ ] `vec_all_nge` - [ ] `vec_all_ngt` - [ ] `vec_and` - [ ] `vec_andc` - [ ] `vec_any_eq` - [ ] `vec_any_ge` - [ ] `vec_any_gt` - [ ] `vec_any_le` - [ ] `vec_any_lt` - [ ] `vec_any_ne` - [ ] `vec_any_nge` - [ ] `vec_any_ngt` - [ ] `vec_avg` - [ ] `vec_blend` - [ ] `vec_bperm` - [ ] `vec_bperm_u128` - [ ] `vec_cmpeq` - [ ] `vec_cmpeq_idx` - [ ] `vec_cmpeq_idx_cc` - [ ] `vec_cmpeq_or_0_idx` - [ ] `vec_cmpeq_or_0_idx_cc` - [ ] `vec_cmpge` - [ ] `vec_cmpgt` - [ ] `vec_cmple` - [ ] `vec_cmplt` - [ ] `vec_cmpne_idx` - [ ] `vec_cmpne_idx_cc` - [ ] `vec_cmpne_or_0_idx` - [ ] `vec_cmpne_or_0_idx_cc` - [ ] `vec_cmpnrg` - [ ] `vec_cmpnrg_cc` - [ ] `vec_cmpnrg_idx` - [ ] `vec_cmpnrg_idx_cc` - [ ] `vec_cmpnrg_or_0_idx` - [ ] `vec_cmpnrg_or_0_idx_cc` - [ ] `vec_cmprg` - [ ] `vec_cmprg_cc` - [ ] `vec_cmprg_idx` - [ ] `vec_cmprg_idx_cc` - [ ] `vec_cmprg_or_0_idx` - [ ] `vec_cmprg_or_0_idx_cc` - [ ] `vec_cntlz` - [ ] `vec_cnttz` - [ ] `vec_cp_until_zero` - [ ] `vec_cp_until_zero_cc` - [ ] `vec_ctd` - [ ] `vec_ctd_s64` - [ ] `vec_ctd_u64` - [ ] `vec_ctsl` - [ ] `vec_ctul` - [ ] `vec_double` - [ ] `vec_eqv` - [ ] `vec_evaluate` - [ ] `vec_extend_s64` - [ ] `vec_extract` - [ ] `vec_find_any_eq` - [ ] `vec_find_any_eq_cc` - [ ] `vec_find_any_eq_idx` - [ ] `vec_find_any_eq_idx_cc` - [ ] `vec_find_any_eq_or_0_idx` - [ ] `vec_find_any_eq_or_0_idx_cc` - [ ] `vec_find_any_ne` - [ ] `vec_find_any_ne_cc` - [ ] `vec_find_any_ne_idx` - [ ] `vec_find_any_ne_idx_cc` - [ ] `vec_find_any_ne_or_0_idx` - [ ] `vec_find_any_ne_or_0_idx_cc` - [ ] `vec_float` - [ ] `vec_fp_test_data_class` - [ ] `vec_gather_element` - [ ] `vec_gfmsum` - [ ] `vec_gfmsum_accum` - [ ] `vec_insert` - [ ] `vec_insert_and_zero` - [ ] `vec_ld2f` - [ ] `vec_load_bndry` - [ ] `vec_load_len` - [ ] `vec_load_len_r` - [ ] `vec_load_pair` - [ ] `vec_madd` - [x] `vec_max` - [ ] `vec_meadd` - [ ] `vec_mergeh` - [ ] `vec_mergel` - [ ] `vec_mhadd` - [x] `vec_min` - [ ] `vec_mladd` - [ ] `vec_moadd` - [ ] `vec_msub` - [ ] `vec_msum` - [ ] `vec_msum_u128` - [ ] `vec_mule` - [ ] `vec_mulh` - [ ] `vec_mulo` - [ ] `vec_nabs` - [ ] `vec_nand` - [ ] `vec_nmadd` - [ ] `vec_nmsub` - [ ] `vec_nor` - [ ] `vec_or` - [ ] `vec_orc` - [ ] `vec_pack` - [ ] `vec_packs` - [ ] `vec_packs_cc` - [ ] `vec_packsu` - [ ] `vec_packsu_cc` - [ ] `vec_perm` - [ ] `vec_permi` - [ ] `vec_popcnt` - [ ] `vec_promote` - [ ] `vec_revb` - [ ] `vec_reve` - [ ] `vec_rl` - [ ] `vec_rl_mask` - [ ] `vec_rli` - [ ] `vec_scatter_element` - [ ] `vec_search_string_cc` - [ ] `vec_search_string_until_zero_cc` - [ ] `vec_sel` - [ ] `vec_signed` - [ ] `vec_slb` - [ ] `vec_sld` - [ ] `vec_sldb` - [ ] `vec_sldw` - [ ] `vec_sll` - [ ] `vec_splat` - [x] `vec_splats` - [ ] `vec_sqrt` - [ ] `vec_srab` - [ ] `vec_sral` - [ ] `vec_srb` - [ ] `vec_srdb` - [ ] `vec_srl` - [ ] `vec_st2f` - [ ] `vec_store_len` - [ ] `vec_store_len_r` - [ ] `vec_sub_u128` - [ ] `vec_subc` - [ ] `vec_subc_u128` - [ ] `vec_sube_u128` - [ ] `vec_subec_u128` - [ ] `vec_sum` - [ ] `vec_sum2` - [ ] `vec_sum4` - [ ] `vec_sum_u128` - [ ] `vec_test_mask` - [ ] `vec_unpackh` - [ ] `vec_unpackl` - [ ] `vec_unsigned` - [ ] `vec_xl` - [ ] `vec_xld2` - [ ] `vec_xlw4` - [ ] `vec_xor` - [ ] `vec_xst` - [ ] `vec_xstd2` - [ ] `vec_xstw4` ### Steps / History <!-- For larger features, more steps might be involved. If the feature is changed later, please add those PRs here as well. --> - [ ] Implementation: #... - [ ] Final comment period (FCP)[^1] - [ ] Stabilization PR <!-- Once the feature has gone through a few release cycles and there are no unresolved questions left, the feature might be ready for stabilization. If this feature didn't go through the RFC process, a final comment period (FCP) is always needed before stabilization. This works as follows: A library API team member can kick off the stabilization process, at which point the rfcbot will ask all the team members to verify they agree with stabilization. Once enough members agree and there are no concerns, the final comment period begins: this issue will be marked as such and will be listed in the next This Week in Rust newsletter. If no blocking concerns are raised in that period of 10 days, a stabilization PR can be opened by anyone. --> ### Unresolved Questions <!-- Include any open questions that need to be answered before the feature can be stabilised. If multiple (unrelated) big questions come up, it can be a good idea to open a separate issue for each, to make it easier to keep track of the discussions. It's useful to link any relevant discussions and conclusions (whether on GitHub, Zulip, or the internals forum) here. --> - None yet. [^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html @rustbot label O-SystemZ general s390x vector/intrinsics progress is tracked at https://github.com/rust-lang/rust/issues/130869 cc @taiki-e
T-libs-api,C-tracking-issue,O-SystemZ
low
Minor
2,796,952,009
tensorflow
Aborted in `tensorflow.raw_ops.RecordInput`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.raw_ops.RecordInput` triggers crash. The crash occurs in the operator RecordInput ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow file_pattern = '@\x0chnFvGe#83FxIIw.\';-$`E?8Y<>-C2p>\x0bl\\B*;z;]a4lrH{gWK}rAY{;T]2k&j-EB)pQu{JLy1K{Q|;g:>CEg<hA"^((UXJ?fMf\'NsZ@KK\rjp{mfE \tYx-3H3;5z\t(%g>\r<!G&_jqx-@o\rz"1 NS"eiM/ 6X/r+/\rj8h>6k{-B]P\n>[\'s0FOZ\x0b[aV\x0c1q\\!{D~2ZIBe{8v9;,oG=Hjsz&~.$T\x0b[_%/ 2;E:Hy8H\tx|dUPg2;[c_\'(>0z-v[Yob\n\x0c;aKX*F.h E@>\x0bK| zGnevJS7G=N_{c}\x0cS4p\']#1`T?FAep' file_random_seed = 0 batch_size = -8494458287779632258 tensorflow.raw_ops.RecordInput(file_pattern=file_pattern, file_random_seed=file_random_seed, batch_size=batch_size) ``` ### Relevant log output ```shell 2025-01-18 12:21:05.765578: F tensorflow/core/framework/tensor_shape.cc:201] Non-OK-status: InitDims(dim_sizes) status: INVALID_ARGUMENT: Expected shape dimensions to be non-negative, got -8494458287779632258 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,953,954
tensorflow
Aborted in `tensorflow.compat.v1.nn.depthwise_conv2d_native`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.compat.v1.nn.depthwise_conv2d_native` triggers crash. The crash occurs in the operator DepthwiseConv2DNative ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = -18261.482131447112 strides = [13762484222233278712, 17895496137899471100] padding = 'VALID' dilations = [11529923499597944069, 6104546329631715214, 16719022861480294065, 15641088296656225516, 5831796171324899039, 2019887057379389965, 6603155403475823832, 128243198633280011, 1962111378253501566, 12787126510677041415, 10754291262680183180] filter = np.array([[ 36832. , -8776. , 255.2, -31200. , -6408. , -12656. , -31696. , -65024. , -50688. ]], dtype=np.float16) tensorflow.compat.v1.nn.depthwise_conv2d_native(input=input, filter=filter, strides=strides, padding=padding, dilations=dilations) ``` ### Relevant log output ```shell 2025-01-18 12:25:34.220155: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2755] Non-OK-status: GetNodeAttr(orig_node->def(), "dilations", &dilations) status: INVALID_ARGUMENT: Attr dilations has value 6104546329631715214 out of range for an int32 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,954,403
electron
Mouse events in unfocused window
### Preflight Checklist - [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success. ### Electron Version 34.0.0 ### What operating system(s) are you using? macOS ### Operating System Version 15.2 ### What arch are you using? arm64 (including Apple Silicon) ### Last Known Working Electron version _No response_ ### Expected Behavior `mouseleave` `mouseout` The timing of these should be the same whether the window has focus or not. ### Actual Behavior When the mouse cursor moves over or off a focused window, it works normally. However, when you move the mouse cursor over an unfocused window, `mouseleave` and `mouseout` occur immediately after `mouseenter` and `mouseover`. And when you move the mouse cursor off, `mouseleave` and `mouseout` do not occur. My environment: - Electron v34.0.0 - macOS 15.2 arm64 ### Testcase Gist URL _No response_ ### Additional Information I created a minimal application to recreate this. I also created a video of the reenactment. https://github.com/sdr0x07b6/electron-mouse-events-test https://github.com/user-attachments/assets/bc13d84f-edb0-4bdc-9a05-c7200edfb531
platform/macOS,bug :beetle:,has-repro-repo,34-x-y
low
Critical
2,796,955,278
tensorflow
Aborted in `tensorflow.raw_ops.ScatterNdNonAliasingAdd`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.raw_ops.ScatterNdNonAliasingAdd` triggers crash. The crash occurs in the operator ScatterNdNonAliasingAdd ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = np.array([False, True, True, False, True, False, True, False, False, False, False, True, True, False, False, True, True]) indices = 1547981256 updates = np.array([ True, False, True, False, True]) tensorflow.raw_ops.ScatterNdNonAliasingAdd(input=input, indices=indices, updates=updates) ``` ### Relevant log output ```shell 2025-01-18 12:28:50.452498: F ./tensorflow/core/framework/tensor.h:847] Check failed: new_num_elements == NumElements() (1 vs. 5) Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,957,488
tensorflow
Aborted in `tensorflow.compat.v1.nn.conv2d_backprop_filter`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.compat.v1.nn.conv2d_backprop_filter` triggers crash. The crash occurs in the operator Conv2DBackpropFilter ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow input = np.array([1.2807603e+38], dtype=np.float32) filter_sizes = np.ones((20, 6, 5, 3), dtype=np.uint16) strides = [4538426903074820908] padding = 'VALID' out_backprop = np.ones((19, 5, 11, 4), dtype=np.float32) tensorflow.compat.v1.nn.conv2d_backprop_filter(input=input, filter_sizes=filter_sizes, out_backprop=out_backprop, strides=strides, padding=padding) ``` ### Relevant log output ```shell 2025-01-18 12:34:39.686592: F tensorflow/core/common_runtime/mkl_layout_pass.cc:2754] Non-OK-status: GetNodeAttr(orig_node->def(), "strides", &strides) status: INVALID_ARGUMENT: Attr strides has value 4538426903074820908 out of range for an int32 Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,958,241
storybook
[Bug]: Next.js-Vite fails in new Next.js project
### Describe the bug #### What I did - `pnpx create next-app` - `pnpx storybook@latest init` - manually replace `@storybook/nextjs` with `@storybook/experimental-nextjs-vite` ![Image](https://github.com/user-attachments/assets/b562bd45-ea01-406b-add8-37740d13968a) #### Workaround - Manually install `styled-jsx` ### Reproduction link - ### Reproduction steps _No response_ ### System ```bash any? ``` ### Additional context _No response_
bug,dependencies,has workaround,nextjs,sev:S2
low
Critical
2,796,960,526
vscode
Git - Add autocomplete for SCM's commit message
Automatically complete commit message based on files in the staging. I believe this is an exceptionally innovative and highly practical feature. <img width="707" alt="Image" src="https://github.com/user-attachments/assets/9709cd24-9486-428d-b9ce-561f8fdf98a8" />
feature-request,git
low
Minor
2,796,962,703
tensorflow
Aborted in `tensorflow.nn.depthwise_conv2d_backprop_filter`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.16.1 tf 2.17.0 ### Custom code Yes ### OS platform and distribution Ubuntu 20.04 ### Mobile device _No response_ ### Python version 3.11 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? On specific inputs, `tensorflow.nn.depthwise_conv2d_backprop_filterr` triggers crash. The crash occurs in the operator DepthwiseConv2dNativeBackpropFilter ### Standalone code to reproduce the issue ```shell import pickle import numpy as np import tensorflow filter_sizes = np.ones((13, 2, 5, 4), dtype=np.uint32) input = 26262.2175547925 out_backprop = -14557.552005335412 strides = [] padding = 'VALID' tensorflow.nn.depthwise_conv2d_backprop_filter(input=input, filter_sizes=filter_sizes, out_backprop=out_backprop, strides=strides, padding=padding) ``` ### Relevant log output ```shell 2025-01-18 12:47:13.553183: F ./tensorflow/core/util/tensor_format.h:428] Check failed: index >= 0 && index < num_total_dims Invalid index from the dimension: 0, 0, N Aborted (core dumped) ```
stat:awaiting response,type:bug,comp:ops,2.17
medium
Critical
2,796,977,129
yt-dlp
[Google Drive] Cannot download source file when cookies are given
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE - [x] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field ### Checklist - [x] I'm reporting that yt-dlp is broken on a **supported** site - [x] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels)) - [x] I've checked that all provided URLs are playable in a browser with the same IP and same login details - [x] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command) - [x] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates - [x] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue) - [x] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required ### Region worldwide ### Provide a description that is worded well enough to be understood When trying to download the source file for a google drive file, I seem to be getting Error 400 when cookies are provided. (example provided below). Now for publicly available files its not much of an issue to run without cookies to get the source file, but for privately shared files, or those on my own drive, I won't be able to access them without cookies, also video files that haven't been processed and don't have any other option other than source. Not sure if the below are related, not knowledgeable enough to tell, but doesn't seem like the exact same issue. [#4580](https://github.com/yt-dlp/yt-dlp/issues/4580) [#9313](https://github.com/yt-dlp/yt-dlp/issues/9313) If you are wondering why I would download files with yt-dlp from google drive, its is because I use MPV, which uses a yt-dlp hook, to stream google drive files. ### Provide verbose output that clearly demonstrates the problem - [x] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`) - [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead - [x] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below ### Complete Verbose Output ```shell */ With cookies /* yt-dlp -vU -F --cookies-from-browser=firefox https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing [debug] Command-line config: ['-vU', '-F', '--cookies-from-browser=firefox', 'https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing'] [debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8 [debug] yt-dlp version [email protected] from yt-dlp/yt-dlp [c8541f8b1] (pip) [debug] Python 3.13.1 (CPython AMD64 64bit) - Windows-10-10.0.19045-SP0 (OpenSSL 3.0.15 3 Sep 2024) [debug] exe versions: ffmpeg 2023-08-28-git-b5273c619d-full_build-www.gyan.dev (setts), ffprobe 2023-08-28-git-b5273c619d-full_build-www.gyan.dev [debug] Optional libraries: Cryptodome-3.21.0, brotli-1.1.0, certifi-2024.08.30, mutagen-1.47.0, requests-2.32.3, sqlite3-3.45.3, urllib3-2.2.3, websockets-14.1 [debug] Proxy map: {} Extracting cookies from firefox [debug] Extracting cookies from: "C:\Users\Name\AppData\Roaming\Mozilla\Firefox\Profiles\4on6caty.default\cookies.sqlite" Extracted 2771 cookies from firefox [debug] Request Handlers: urllib, requests, websockets [debug] Loaded 1837 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest Latest version: [email protected] from yt-dlp/yt-dlp yt-dlp is up to date ([email protected] from yt-dlp/yt-dlp) [GoogleDrive] Extracting URL: https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing [GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Downloading video webpage [GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting source file [GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting confirmed source file WARNING: [GoogleDrive] Unable to request confirmed source file: HTTP Error 400: Bad Request [debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, hdr:12(7), vcodec, channels, acodec, size, br, asr, proto, vext, aext, hasaud, source, id [info] Available formats for 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: ID EXT RESOLUTION │ PROTO │ VCODEC ACODEC ─────────────────────────────────────────── 18 mp4 640x360 │ https │ unknown unknown 22 mp4 1280x720 │ https │ unknown unknown 37 mp4 1920x1080 │ https │ unknown unknown */ without cookies /* yt-dlp -vU -F https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing [debug] Command-line config: ['-vU', '-F', 'https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing'] [debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8 [debug] yt-dlp version [email protected] from yt-dlp/yt-dlp [c8541f8b1] (pip) [debug] Python 3.13.1 (CPython AMD64 64bit) - Windows-10-10.0.19045-SP0 (OpenSSL 3.0.15 3 Sep 2024) [debug] exe versions: ffmpeg 2023-08-28-git-b5273c619d-full_build-www.gyan.dev (setts), ffprobe 2023-08-28-git-b5273c619d-full_build-www.gyan.dev [debug] Optional libraries: Cryptodome-3.21.0, brotli-1.1.0, certifi-2024.08.30, mutagen-1.47.0, requests-2.32.3, sqlite3-3.45.3, urllib3-2.2.3, websockets-14.1 [debug] Proxy map: {} [debug] Request Handlers: urllib, requests, websockets [debug] Loaded 1837 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest Latest version: [email protected] from yt-dlp/yt-dlp yt-dlp is up to date ([email protected] from yt-dlp/yt-dlp) [GoogleDrive] Extracting URL: https://drive.google.com/file/d/1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-/view?usp=sharing [GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Downloading video webpage [GoogleDrive] 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: Requesting source file [debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, hdr:12(7), vcodec, channels, acodec, size, br, asr, proto, vext, aext, hasaud, source, id [info] Available formats for 1SlteKbVi2xOTzFEcGoyTNnoN8gcrW8v-: ID EXT RESOLUTION │ PROTO │ VCODEC ACODEC ─────────────────────────────────────────────── 18 mp4 640x360 │ https │ unknown unknown 22 mp4 1280x720 │ https │ unknown unknown 37 mp4 1920x1080 │ https │ unknown unknown source mp4 unknown │ https │ unknown unknown ```
account-needed,site-bug,triage,can-share-account
low
Critical
2,796,983,161
flutter
[go_router] App link provided to an app is ignored by GoRouter, when the app is suspended and resumed from background (Android)
### Steps to reproduce Hello, I have an app with native GoRouter restoration implemented (by providing restorationScopeId). When I provide an app link to my app, which is in the background and suspended by the OS to save memory, GoRouter ignores the app link and restores the route from its previous state. However, when the app is not started, in the foreground, or not suspended, the app link works as intended. GoRouter successfully opens the route provided by the link. Can you please advise on the issue? Thank you! ### Expected results GoRouter opens the route according to the app link when an app link is provided to an app that is suspended in the background. ### Actual results When an app link is provided to an app that is suspended in the background, GoRouter ignores the app link and restores the last route. ### Code sample I don't think that providing a code sample is relevant here ### Screenshots or Video ### Logs ### Flutter Doctor output flutter doctor -v [✓] Flutter (Channel stable, 3.27.2, on Linux Mint 21.3 5.15.0-130-generic, locale en_US.UTF-8) • Flutter version 3.27.2 on channel stable at /home/ruslan/fvm/versions/3.27.2 • Upstream repository ssh://[email protected]/flutter/flutter.git • Framework revision 68415ad1d9 (5 days ago), 2025-01-13 10:22:03 -0800 • Engine revision e672b006cb • Dart version 3.6.1 • DevTools version 2.40.2
platform-android,package,has reproducible steps,p: go_router,team-go_router,found in release: 3.27,found in release: 3.28
low
Major
2,796,983,732
ollama
Deepseek-v3 make smaller quants available
Hi, The smallest Deepseek-v3 you currently offer via the library is deepseek-v3:671b-q4_K_M. Please can you make smaller quants available, e.g., Q3, to make this model more accessible. Thanks!
model request
low
Minor
2,796,984,999
transformers
Inconsistent output lengths when `max_length=20` is set implicitly vs explicitly in `generate()`
### System Info - `transformers` version: 4.49.0.dev0 - Platform: macOS-15.1.1-arm64-arm-64bit - Python version: 3.11.10 - Huggingface_hub version: 0.27.1 - Safetensors version: 0.5.2 - Accelerate version: 1.2.1 - Accelerate config: not found - PyTorch version (GPU?): 2.5.1 (False) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: no ### Who can help? @gante @ArthurZucker Related PR that discusses recent default `max_length`-related changes: #34814. ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction When using `generate()` with a model that has `generation_config.max_length=20`, the output length differs depending on whether `max_length` is passed explicitly or used implicitly from the `generation_config`. ```python from transformers import AutoModelForCausalLM, AutoTokenizer # Setup from tests/generation/test_utils.py::GenerationIntegrationTests article = "Today a dragon flew over Paris." model = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2") tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2") input_ids = tokenizer(article, return_tensors="pt").input_ids # Case 1: Implicit max_length from generation_config out_gen_implicit = model.generate(input_ids=input_ids) print(out_gen_implicit.shape[-1]) # 36 # Case 2: Explicit max_length out_gen_explicit = model.generate( input_ids=input_ids, max_length=model.generation_config.max_length ) print(out_gen_explicit.shape[-1]) # 20 ``` In the first case, the generated text is longer than in the second case (36 vs. 20 tokens). #### Reason and scope In the first case, `max_length` is overwritten as follows in file `src/transformers/generation/utils.py`, function `_prepare_generated_length`: ```python if generation_config.max_length == GenerationConfig().max_length: generation_config.max_length = generation_config.max_length + input_ids_length ``` Since `GenerationConfig().max_length` defaults to 20, the bug only affects models with `generation_config.max_length` set to 20. ### Expected behavior The calls `model.generate(input_ids=input_ids)` and `model.generate(input_ids=input_ids, max_length=model.generation_config.max_length)` should generate texts of the same length when `generation_config.max_length` is set to 20.
bug
low
Critical
2,796,992,863
godot
[4.4] Camera3D preview doesn't show anything until interact with editor.
### Tested versions v4.4.beta1.official [d33da79d3] ### System information Godot v4.4.beta1 - Android - Single-window, 1 monitor - OpenGL ES 3 (Compatibility) - Adreno (TM) 610 - (8 threads) ### Issue description After add Camera3D inside a scene camera preview doesn't show anything. It show nothing until I interact with the editor. Also this happen when select the Camera3D node every time. Video: https://github.com/user-attachments/assets/37caac9f-d2f8-44b8-8086-d9fb79f80c4a ### Steps to reproduce - add camera3D inside your tree - check the preview ### Minimal reproduction project (MRP) n/a
bug,topic:editor,topic:3d
low
Minor
2,797,019,519
flutter
Grade Plugin Flutter
### Steps to reproduce flutter build appbundle FAILURE: Build failed with an exception. * What went wrong: A problem occurred configuring root project 'android'. > Could not resolve all files for configuration ':classpath'. > Could not find io.flutter:flutter-gradle-plugin:3.0.0. Searched in the following locations: - https://dl.google.com/dl/android/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom - https://repo.maven.apache.org/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom - https://storage.googleapis.com/download.flutter.io/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom Required by: project : * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. > Get more help at https://help.gradle.org. BUILD FAILED in 4s Running Gradle task 'bundleRelease'... 5,2s Gradle task bundleRelease failed with exit code 1 ### Actual results I am trying to build the appbundle file, but I get these errors ### Logs C:\Users\GGS\vocabulary>flutter pub outdated Showing outdated packages. [*] indicates versions that are not the latest available. Package Name Current Upgradable Resolvable Latest direct dependencies: all up-to-date. dev_dependencies: all up-to-date. transitive dev_dependencies: boolean_selector *2.1.1 *2.1.1 *2.1.1 2.1.2 fake_async *1.3.1 *1.3.1 *1.3.1 1.3.2 leak_tracker *10.0.7 *10.0.7 *10.0.7 10.0.8 leak_tracker_flutter_testing *3.0.8 *3.0.8 *3.0.8 3.0.9 matcher *0.12.16+1 *0.12.16+1 *0.12.16+1 0.12.17 stack_trace *1.12.0 *1.12.0 *1.12.0 1.12.1 stream_channel *2.1.2 *2.1.2 *2.1.2 2.1.4 test_api *0.7.3 *0.7.3 *0.7.3 0.7.4 vm_service *14.3.0 *14.3.0 *14.3.0 15.0.0 all dependencies are up-to-date. C:\Users\GGS\vocabulary>flutter clean C:\Users\GGS\vocabulary>flutter pub get Resolving dependencies... Downloading packages... ! async 2.12.0 (overridden) boolean_selector 2.1.1 (2.1.2 available) ! characters 1.4.0 (overridden) ! clock 1.1.2 (overridden) ! collection 1.19.1 (overridden) ! csslib 1.0.2 (overridden) fake_async 1.3.1 (1.3.2 available) ! html 0.15.5 (overridden) leak_tracker 10.0.7 (10.0.8 available) leak_tracker_flutter_testing 3.0.8 (3.0.9 available) matcher 0.12.16+1 (0.12.17 available) ! material_color_utilities 0.12.0 (overridden) ! meta 1.16.0 (overridden) ! path 1.9.1 (overridden) ! petitparser 6.1.0 (overridden) ! source_span 1.10.1 (overridden) stack_trace 1.12.0 (1.12.1 available) stream_channel 2.1.2 (2.1.4 available) ! string_scanner 1.4.1 (overridden) ! term_glyph 1.2.2 (overridden) test_api 0.7.3 (0.7.4 available) vm_service 14.3.0 (15.0.0 available) Got dependencies! 9 packages have newer versions incompatible with dependency constraints. Try `flutter pub outdated` for more information. C:\Users\GGS\vocabulary>flutter build appbundle FAILURE: Build failed with an exception. * What went wrong: A problem occurred configuring root project 'android'. > Could not resolve all files for configuration ':classpath'. > Could not find io.flutter:flutter-gradle-plugin:3.0.0. Searched in the following locations: - https://dl.google.com/dl/android/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom - https://repo.maven.apache.org/maven2/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom - https://storage.googleapis.com/download.flutter.io/io/flutter/flutter-gradle-plugin/3.0.0/flutter-gradle-plugin-3.0.0.pom Required by: project : * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. > Get more help at https://help.gradle.org. BUILD FAILED in 4s Running Gradle task 'bundleRelease'... 5,2s Gradle task bundleRelease failed with exit code 1 ### Flutter Doctor output flutter doctor -v [√] Flutter (Channel stable, 3.27.2, on Microsoft Windows [Versione 10.0.19045.5371], locale it-IT) • Flutter version 3.27.2 on channel stable at C:\flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 68415ad1d9 (5 days ago), 2025-01-13 10:22:03 -0800 • Engine revision e672b006cb • Dart version 3.6.1 • DevTools version 2.40.2 [√] Windows Version (Installed version of Windows is version 10 or higher) [√] Android toolchain - develop for Android devices (Android SDK version 34.0.0) • Android SDK at C:\Users\GGS\AppData\Local\Android\Sdk • Platform android-34, build-tools 34.0.0 • Java binary at: C:\Program Files\Android\Android Studio\jbr\bin\java • Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-10027231) • All Android licenses accepted. [√] Chrome - develop for the web • Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe [X] Visual Studio - develop Windows apps X Visual Studio not installed; this is necessary to develop Windows apps. Download at https://visualstudio.microsoft.com/downloads/. Please install the "Desktop development with C++" workload, including all of its default components [√] Android Studio (version 2022.3) • Android Studio at C:\Program Files\Android\Android Studio • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-10027231) [√] Connected device (3 available) • Windows (desktop) • windows • windows-x64 • Microsoft Windows [Versione 10.0.19045.5371] • Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83 • Edge (web) • edge • web-javascript • Microsoft Edge 131.0.2903.146 [√] Network resources • All expected network resources are available. ! Doctor found issues in 1 category.
waiting for customer response,in triage
low
Critical
2,797,020,257
godot
GPUParticles3D nodes skip update for one frame whenever a new GPUParticles3D is created
### Tested versions Reproducible in: v4.4.beta1.mono.official [d33da79d3] ### System information Godot v4.4.beta1.mono - Windows 10 (build 19045) - Single-window, 1 monitor - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2060 (NVIDIA; 31.0.15.5275) - Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz (12 threads) ### Issue description Whenever a scene with a GPUParticles3D node is instantiated, all existing GPUParticles3D nodes skip updates for one frame. Adding the instantiated scene to the scene tree via e.g. `add_sibling` is not necessary for the issue to occur. The issue is difficult to see with high FPS and/or slow moving particles, but easy to notice with a low FPS cap (e.g. 20) and fast moving particles. https://github.com/user-attachments/assets/56202ffb-1775-4385-9c0a-3f830713a3ae The GIF below illustrates the moment of creating a new particle system in slowdown. It can be seen that the frame following the one with the creation of the new GPUParticles3D has the exact same configuration of all particles as the frame prior, but slightly offset due to independent camera movement. ![Image](https://github.com/user-attachments/assets/42345780-e434-4434-94bd-915a90206b7e) ### Steps to reproduce - In the project settings, set the max_fps option to a low value, e.g. 20. This will make the issue easier to see. - Create a scene with a GPUParticles3D as a root node. The content is unimportant as long as the particles are visible and not one-shot. - Create a preview scene. - Put a Camera3D node in the root of the preview scene, e.g. unrotated, at z=6m. This is for preview purposes. - Put a Node3D node in the root of the preview scene, e.g. unrotated, at x=-7, y=5. This is a cursor for spawning the particle scene. Attach a script to the node, for example: ```gdscript extends Node3D @export var particle_scene: PackedScene @onready var spawned_particles: Array = [] func _physics_process(_delta): if Input.is_action_just_pressed("ui_accept"): if global_position.x < 7.0: # create a new particle system var particle = particle_scene.instantiate() particle.position = global_position add_sibling(particle) spawned_particles.push_back(particle) # push the position global_position.x += 1.0 else: # if the position was pushed far enough, reset for particle in spawned_particles: particle.queue_free() spawned_particles.clear() global_position.x = -7.0 ``` - Make sure the particle scene is linked the the cursor node's export variable. - Run the game in play mode on the preview scene. Pressing Enter should result in the cursor spawning a particle scene and moving. With the low FPS value, it should be easy to see that every time a particle scene is spawned, all the GPUParticles3D nodes stop updating for exactly one frame. Commenting out the `add_sibling` line shows that the issue still occurs when the new GPUParticles3D node is not added to the scene tree. If there are other moving nodes on the scene, it can be seen that this is a local issue rather than a global hang-up. Minimal reproduction project comes with slight camera movement to illustrate that the movement of the rest of the game continues smoothly. ### Minimal reproduction project (MRP) [particletest2_2025-01-18_15-33-43.zip](https://github.com/user-attachments/files/18464787/particletest2_2025-01-18_15-33-43.zip)
bug,confirmed,topic:particles
low
Major
2,797,025,983
vscode
properly not working
Type: <b>Bug</b> work space is not working in this visual studio code VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Windows_NT x64 6.2.9200 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i5-7300U CPU @ 2.60GHz (4 x 2712)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|undefined| |Memory (System)|7.88GB (1.42GB free)| |Process Argv|--crash-reporter-id 678aba3c-fe94-4076-af47-bbf5e066d80f| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (29)</summary> Extension|Author (truncated)|Version ---|---|--- python-debug-address-prompt|ale|1.0.1 prettier-vscode|esb|11.0.0 c-cpp-runner|fra|9.4.9 codespaces|Git|1.17.3 cpptask|kay|0.0.1 csdevkit|ms-|1.15.34 csharp|ms-|2.61.28 vscode-dotnet-runtime|ms-|2.2.5 debugpy|ms-|2024.14.0 python|ms-|2024.22.2 vscode-pylance|ms-|2024.12.1 jupyter|ms-|2024.11.0 jupyter-keymap|ms-|1.1.2 jupyter-renderers|ms-|1.0.21 vscode-jupyter-cell-tags|ms-|0.1.9 vscode-jupyter-slideshow|ms-|0.1.6 cmake-tools|ms-|1.19.52 cpptools|ms-|1.22.11 cpptools-extension-pack|ms-|1.3.0 powershell|ms-|2024.4.0 remote-explorer|ms-|0.4.3 remote-server|ms-|1.5.2 java|red|1.38.0 LiveServer|rit|5.7.9 cmake|twx|0.0.17 vscode-lldb|vad|1.11.1 vscode-gradle|vsc|3.16.2024111106 vscode-java-debug|vsc|0.58.1 vscode-java-test|vsc|0.43.0 (1 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vspor879:30202332 vspor708:30202333 vspor363:30204092 vscod805:30301674 binariesv615:30325510 vsaa593cf:30376535 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 2i9eh265:30646982 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 h48ei257:31000450 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupytercf:31046870 newcmakeconfigv2:31071590 nativerepl1:31139838 pythonrstrctxt:31112756 nativeloc2:31192216 cf971741:31144450 iacca1:31171482 notype1cf:31157160 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 cf1a2727:31215809 6074i472:31201624 dwoutputs:31217127 ``` </details> <!-- generated by issue reporter -->
info-needed,triage-needed
low
Critical
2,797,036,148
rust
Missing re-export of `FromCoroutine` in `core::iter` for `from_coroutine`
The `FromCoroutine` type returned by the unstable function `core::iter::from_coroutine` was not publicly re-exported in the `core::iter` module.
C-bug,T-libs
low
Minor
2,797,037,301
svelte
Cannot set properties of null (setting 'currentTime'), when transition
### Describe the bug When playing audio HTML element with binded currentTime value becomes hidden with transition, the Svelte still tries to set the value of the now missing property (while animated), which results in an error Cannot set properties of null. ### Reproduction Full example: [REPL](https://svelte.dev/playground/6b0bf9775fe64b0097f0fe18d5eee5ad?version=5.19.0) Press play and then hide the player ### Logs ```shell ``` ### System Info ```shell Chrome 131 ``` ### Severity annoyance
transition/animation
low
Critical
2,797,039,519
transformers
Defining LLM Dataset types in Trainers or during Training Workflow
### Feature request Hi, Just wondering if its possible to define a way to define the dataset input labels mapping argument in the transformers.Trainer or something similar? I understand that the input data labels are dependent on the type of training loss function used and problem for example, the type of defining attention mask ids for Mask learner vs Causal learners etc. Would also be extremely helpful if I can have suggestions on how I can check what dataset column labels the trainer/model will read? -- besides the `help(model.forward)` as this really just shows the model's inputs as it is into the learner. Any suggestions would be greatly appreciated! thanks! ### Motivation This problem is giving me clinical depression ### Your contribution If there's any suggestions in solving this case, I would love to contribute
Feature request
low
Minor
2,797,048,027
langchain
OpenAI Assistant V2 Runnable parameters, run_metadata does not exist
### Checked other resources - [x] I added a very descriptive title to this issue. - [x] I searched the LangChain documentation with the integrated search. - [x] I used the GitHub search to find a similar question and didn't find it. - [x] I am sure that this is a bug in LangChain rather than my code. - [x] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code ```python from langchain_community.agents.openai_assistant import OpenAIAssistantV2Runnable assistant = OpenAIAssistantV2Runnable.create_assistant( name="bug_reproducer", instructions='Write "A"', check_every_ms=5_000.0, tools={}, model="gpt-4o-mini", ) message = { "content": "A", "max_completion_tokens": 10, "run_metadata": {"key": "value"}, } resp = assistant.invoke(message) print(resp) ``` ### Error Message and Stack Trace (if applicable) Traceback (most recent call last): File "[Redacted]/Langchain Tests/reproducer.py", line 17, in <module> resp = assistant.invoke(message) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "[Redacted]/lib/python3.12/site-packages/langchain_community/agents/openai_assistant/base.py", line 383, in invoke raise e File "[Redacted]/langchain_community/agents/openai_assistant/base.py", line 365, in invoke run = self._create_thread_and_run(input, thread) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "[Redacted]/langchain_community/agents/openai_assistant/base.py", line 579, in _create_thread_and_run run = self.client.beta.threads.create_and_run( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "[Redacted]/openai/_utils/_utils.py", line 279, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ TypeError: Threads.create_and_run() got an unexpected keyword argument 'run_metadata' ### Description Currently the parameters passed down to OpenAI from Langchain are limited to a set of parameters (lately updated in this [PR](https://github.com/langchain-ai/langchain/pull/29207) ). While testing the usage of the parameters I noticed that one of the parameters already present prior to the PR that could be passed down to openAI is "run_metadata", but verifying on OpenAI library I saw no such parameters, a parameter "metadata" [exists](https://github.com/openai/openai-python/blob/main/src/openai/resources/beta/threads/threads.py#L275) but not the "run_metadata" one. Going back to [OpenAI 1.1.0](https://github.com/openai/openai-python/blob/v1.1.0/src/openai/resources/beta/threads/threads.py#L201) which introduced assistants I always see "metadata" and not "run_metadata", so that parameter if passed never actually worked). This could be easily solved by changing the occurences of the parameter "run_metadata" with "metadata" in the allow sets. But given the current status of the Assistant class and the passage of parameters to OpenAI threads and runs underneath, I was thinking that the current approach with a "set of allowed values" is quite strict, can be edited only via a PR and could break as soon as a change is done on OpenAI python package. To avoid this kind of issues and keep aligned parameters from OpenAI side and Langchain side the parameters should be passed in a separate object (not in additional keys of the input where we also provide the chat messages) and could be useful to somehow use for example for `thread_create_and_run` method to use the "TypedDict" [`ThreadCreateAndRunParamsBase`](https://github.com/openai/openai-python/blob/main/src/openai/types/beta/thread_create_and_run_params.py#L38) present in OpenAI library. ### System Info python 3.12 pip freeze aiohappyeyeballs==2.4.4 aiohttp==3.11.11 aiosignal==1.3.2 annotated-types==0.7.0 anyio==4.8.0 attrs==24.3.0 certifi==2024.12.14 charset-normalizer==3.4.1 dataclasses-json==0.6.7 distro==1.9.0 frozenlist==1.5.0 greenlet==3.1.1 h11==0.14.0 httpcore==1.0.7 httpx==0.28.1 httpx-sse==0.4.0 idna==3.10 jiter==0.8.2 jsonpatch==1.33 jsonpointer==3.0.0 langchain==0.3.14 langchain-community==0.3.14 langchain-core==0.3.30 langchain-text-splitters==0.3.5 langsmith==0.2.11 marshmallow==3.25.1 multidict==6.1.0 mypy-extensions==1.0.0 numpy==2.2.1 openai==1.59.8 orjson==3.10.14 packaging==24.2 propcache==0.2.1 pydantic==2.10.5 pydantic-settings==2.7.1 pydantic_core==2.27.2 python-dotenv==1.0.1 PyYAML==6.0.2 requests==2.32.3 requests-toolbelt==1.0.0 sniffio==1.3.1 SQLAlchemy==2.0.37 tenacity==9.0.0 tqdm==4.67.1 typing-inspect==0.9.0 typing_extensions==4.12.2 urllib3==2.3.0 yarl==1.18.3
🤖:bug
low
Critical
2,797,056,331
pytorch
Driver Allocated Memory grows unrestricted when using torch.unique on MPS device
### 🐛 Describe the bug When using `torch.unique` in a loop on the MPS backend, the memory allocated by the driver grows unrestricted. In my real application that leads to an `RuntimeError: MPS backend out of memory (MPS allocated: 24.00 MB, other allocations: 36.24 GB, max allowed: 36.27 GB)` error late in the training. I created this minimal example with the same behaviour. ```python import torch import gc def test_operations(iterations: int, shape: tuple[int, int]) -> None: print(f"PyTorch version: {torch.__version__}") # Test 1: torch.unique print("\nTest 1: torch.unique") x = torch.randint(0, 2, shape, device="mps") for i in range(iterations): y = torch.unique(x) del y # Empty cache and collect garbage to make sure torch.mps.empty_cache() gc.collect() if i % 10 == 0: print( f"Iter {i}: Driver Allocated Memory: {torch.mps.driver_allocated_memory() / (1024**2):.2f}MB, Current Allocated Memory: {torch.mps.current_allocated_memory() / (1024**2):.2f}MB" ) # Test 2: torch.sort (comparison) print("\nTest 2: torch.sort") for i in range(iterations): y = torch.sort(x)[0] del y # Empty cache and collect garbage to make sure torch.mps.empty_cache() gc.collect() if i % 10 == 0: print( f"Iter {i}: Driver memory: {torch.mps.driver_allocated_memory() / (1024**2):.2f}MB, Current memory: {torch.mps.current_allocated_memory() / (1024**2):.2f}MB" ) test_operations(iterations=100, shape=(2000, 10)) ``` Results in ``` PyTorch version: 2.5.1 Test 1: torch.unique Iter 0: Driver Allocated Memory: 18.73MB, Current Allocated Memory: 0.15MB Iter 10: Driver Allocated Memory: 98.73MB, Current Allocated Memory: 0.15MB Iter 20: Driver Allocated Memory: 178.73MB, Current Allocated Memory: 0.15MB Iter 30: Driver Allocated Memory: 258.73MB, Current Allocated Memory: 0.15MB Iter 40: Driver Allocated Memory: 338.73MB, Current Allocated Memory: 0.15MB Iter 50: Driver Allocated Memory: 418.73MB, Current Allocated Memory: 0.15MB Iter 60: Driver Allocated Memory: 578.73MB, Current Allocated Memory: 0.15MB Iter 70: Driver Allocated Memory: 738.72MB, Current Allocated Memory: 0.15MB Iter 80: Driver Allocated Memory: 898.72MB, Current Allocated Memory: 0.15MB Iter 90: Driver Allocated Memory: 1058.72MB, Current Allocated Memory: 0.15MB Test 2: torch.sort Iter 0: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 10: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 20: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 30: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 40: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 50: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 60: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 70: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 80: Driver memory: 1202.72MB, Current memory: 0.15MB Iter 90: Driver memory: 1202.72MB, Current memory: 0.15MB ``` Showing the increase in the driver allocated memory when using `torch.unique` but not when using another function like `torch.sort`. I just used `torch.sort` as comparison here. Is this behaviour expected? ### Versions Collecting environment information... PyTorch version: 2.5.1 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: macOS 15.2 (arm64) GCC version: Could not collect Clang version: 16.0.0 (clang-1600.0.26.6) CMake version: Could not collect Libc version: N/A Python version: 3.11.11 (main, Dec 3 2024, 17:20:40) [Clang 16.0.0 (clang-1600.0.26.4)] (64-bit runtime) Python platform: macOS-15.2-arm64-arm-64bit Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Apple M2 Pro Versions of relevant libraries: [pip3] torch==2.5.1 [conda] Could not collect cc @kulinseth @albanD @malfet @DenisVieriu97 @jhavukainen
module: memory usage,triaged,module: mps
low
Critical
2,797,056,878
rust
Tracking issue for release notes of #133925: disallow `repr()` on invalid items
This issue tracks the release notes text for #133925. ### Steps - [ ] Proposed text is drafted by PR author (or team) making the noteworthy change. - [ ] Issue is nominated for release team review of clarity for wider audience. - [ ] Release team includes text in release notes/blog posts. ### Release notes text The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing). ````markdown # Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...) - [disallow `repr()` on invalid items](https://github.com/rust-lang/rust/pull/133925) ```` > [!TIP] > Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use. > The category will be de-duplicated with all the other ones by the release team. > > *More than one section can be included if needed.* ### Release blog section If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section. *Otherwise leave it empty.* ````markdown ```` cc @folkertdev, @compiler-errors -- origin issue/PR authors and assignees for starting to draft text
T-lang,relnotes,needs-triage,relnotes-tracking-issue
low
Critical
2,797,065,765
rust
Tracking issue for release notes of #133889: Consider fields to be inhabited if they are unstable
This issue tracks the release notes text for #133889. ### Steps - [ ] Proposed text is drafted by PR author (or team) making the noteworthy change. - [ ] Issue is nominated for release team review of clarity for wider audience. - [ ] Release team includes text in release notes/blog posts. ### Release notes text The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing). ````markdown # Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...) - [Consider fields to be inhabited if they are unstable](https://github.com/rust-lang/rust/pull/133889) ```` > [!TIP] > Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use. > The category will be de-duplicated with all the other ones by the release team. > > *More than one section can be included if needed.* ### Release blog section If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section. *Otherwise leave it empty.* ````markdown ```` cc @compiler-errors, @Nadrieril -- origin issue/PR authors and assignees for starting to draft text
T-lang,relnotes,needs-triage,relnotes-tracking-issue
low
Critical