id
int64
393k
2.82B
repo
stringclasses
68 values
title
stringlengths
1
936
body
stringlengths
0
256k
โŒ€
labels
stringlengths
2
508
priority
stringclasses
3 values
severity
stringclasses
3 values
2,674,581,831
flutter
fllutter create should initialize analytics_options.yaml with stricter settings
### Use case When creating a new flutter project with `fluter create` the generated `analysis_options.yaml` does only contain `flutter_lints` ### Proposal It should at least set ```yaml analyzer: language: strict-casts: true strict-inference: true strict-raw-types: true ``` also some lints should be set to level `errror` see https://github.com/dart-lang/sdk/issues/57098#issuecomment-2485661859
c: new feature,package,c: proposal,p: flutter_lints
low
Minor
2,674,610,621
storybook
Luminosity contrast ratio of keyboard focus indicator with the grey background on the left navigation panel is less than minimum required ratio of 3:1.
### Describe the bug Luminosity contrast ratio of keyboard focus indicator with the grey background on the left navigation panel is less than minimum required ratio of 3:1 **Observation:** Same issue is repro'ing on both Mac and Windows. **User Impact:** Visual impaired users and low vision users will be impacted if Luminosity contrast ratio of keyboard focus indicator is less than the required ratio due to which they will not be able to identify their focus properly and will not be able to complete their task easily. Actual Result: Luminosity contrast ratio of keyboard focus indicator with the grey background for 'Content' Control in the left navigation panel is 1.2:1 which is less than minimum required ratio of 3:1. Refer Attachment: Luminosity contrast ratio of keyboard focus indicator with the grey background is less than minimum required. ![Image](https://github.com/user-attachments/assets/4f0121e6-b7d3-417f-bfaf-980b1b10f885) Expected Result: Luminosity contrast ratio of keyboard focus indicator with the grey background for all the controls in the left navigation panel should be greater than minimum required ratio of 3:1. ### Reproduction link About / Introduction - Page โ‹… Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net) ### Reproduction steps 1. Open URL: [About / Introduction - Page โ‹… Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net)](https://ambitious-cliff-0c8148010.2.azurestaticapps.net/?path=/story/about-introduction--page) in latest edge browser. 2. Press tab key to reach to the controls in the left pane. 3. Press tab key and reach to the 'Content' control. 4. Open Color contrast analyzer and select foreground and background for the focus indicator using color picker. 5. Verify whether on pressing tab keys to reach to the 'Contents' control present in the left navigation panel, luminosity contrast ratio of keyboard focus indicator with the grey background is greater than minimum required or 3:1 or not. ### System ```bash Test Environment: OS: Windows 11 Version 24H2 OS build 26100.1742 Browser: Microsoft New Edge Version129.0.2792.52(Official build) (64-bit) URL: About / Introduction - Page โ‹… Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net) Tool Used: Color Contrast Analyzer Mac Test Environment: OS: macOS Sequoia 15.0.1 Browser: Microsoft New Edge Version130.0.2849.56 (Official build) (64-bit) RL: About / Introduction - Page โ‹… Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net) Tool Used: Color Contrast Analyzer ``` ### Additional context _No response_
bug,accessibility
low
Critical
2,674,611,675
pytorch
DISABLED test_sync (__main__.TestMPS)
Platforms: mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_sync&suite=TestMPS&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/33237372365). Over the past 3 hours, it has been determined flaky in 4 workflow(s) with 10 failures and 4 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_sync` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_mps.py` cc @clee2000 @malfet @albanD @kulinseth @DenisVieriu97 @jhavukainen
triaged,module: flaky-tests,module: macos,skipped,module: mps
low
Critical
2,674,675,963
tensorflow
Aborted (core dumped) in `tf.raw_ops.MatrixSolve`
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tf 2.17 ### Custom code Yes ### OS platform and distribution Linux Ubuntu 22.04.3 LTS (x86_64) ### Mobile device _No response_ ### Python version 3.9.13 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? When the shape of the matrix argument is empty and the gpu is available, tf.raw_ops.MatrixSolve triggers a crash. It can be reproduced on tf-nightly when the gpu is available. ### Standalone code to reproduce the issue ```shell import tensorflow as tf tf.raw_ops.MatrixSolve(matrix=tf.random.uniform([], dtype=tf.dtypes.double, maxval=1000000000), rhs=tf.random.uniform([1, 2], dtype=tf.dtypes.double, maxval=1000000000), adjoint=True) ``` ### Relevant log output ```shell 2024-11-20 14:51:08.714846: I tensorflow/core/util/port.cc:153] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`. 2024-11-20 14:51:08.775383: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:485] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered 2024-11-20 14:51:08.852267: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:8454] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered 2024-11-20 14:51:08.876168: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1452] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered 2024-11-20 14:51:08.931206: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations. To enable the following instructions: SSE4.1 SSE4.2 AVX AVX2 AVX512F AVX512_VNNI AVX512_BF16 AVX512_FP16 AVX_VNNI AMX_TILE AMX_INT8 AMX_BF16 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags. 2024-11-20 14:51:16.385650: I tensorflow/core/common_runtime/gpu/gpu_device.cc:2021] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 535 MB memory: -> device: 0, name: NVIDIA GeForce RTX 4090, pci bus id: 0000:1f:00.0, compute capability: 8.9 2024-11-20 14:51:16.387914: I tensorflow/core/common_runtime/gpu/gpu_device.cc:2021] Created device /job:localhost/replica:0/task:0/device:GPU:1 with 71 MB memory: -> device: 1, name: NVIDIA GeForce RTX 4090, pci bus id: 0000:d4:00.0, compute capability: 8.9 2024-11-20 14:51:16.542383: F tensorflow/core/framework/tensor_shape.cc:356] Check failed: d >= 0 (0 vs. -1) Aborted (core dumped) ```
stat:awaiting tensorflower,type:bug,comp:ops,2.17
medium
Critical
2,674,687,480
pytorch
[AOTInductor] AOT compiled models do not consistently check input shapes
### ๐Ÿ› Describe the bug AOT compiled models do not consistently provide the same runtime input Tensor checks and safety as do either plain PyTorch models or PyTorch exported graph models (via `torch.export`). Oddly, whether these checks happen with AOT models seems to depend on the shape of the data used to export the PyTorch model (???). In the repro below, I export a toy model as a graph and then as an AOT compiled model and run inference on both. In the first case, I seed the "happy path" model with 2D input and both the torch.export'ed and AOT compiled models expectedly raise `RuntimeException`s and complain about input Tensor shape mismatch. However, in the second case, the "sad path", the model is exported with 1D example data and the AOT compiled model **does not** error on invalid input data. Inference completes successfully which is **not** the expected behavior -- I'd expected the AOT compiled model to raise a RuntimeException as it did in the "happy path". **repro** ```python import torch class Model(torch.nn.Module): def __init__(self) -> None: super().__init__() self._layer = torch.nn.Linear(2, 1) def forward(self, X: torch.Tensor) -> torch.Tensor: return self._layer(X) model = Model() # HAPPY PATH: 2D example data print("-" * 50 + "\nHAPPY PATH") with torch.inference_mode(): device = "cuda:0" # device = "cpu" X = torch.rand((1, 2), device=device) model = model.to(device) # AOTI compile on CPU example_inputs = X exported_program = torch.export.export(model, (example_inputs,),) exported_model = exported_program.module().to(device) so_path = torch._inductor.aot_compile(exported_model, (example_inputs,),) aot_model = torch._export.aot_load(so_path, device) # Inference: OK print(">> OK: Exported Graph Model: ", exported_model(X)) print(">> OK: AOT Compiled Model: ", aot_model(X)) # Inference: FAILS # NOTE: shape of X_t is invalid X_t = torch.rand((3, 1), device=device) try: exported_model(X_t) except RuntimeError as exc: print(">> FAIL: Exported Graph Model: ", exc) try: aot_model(X_t) except RuntimeError as exc: print(">> FAIL: AOT Compiled Model: ", exc) print("-" * 50) print("-" * 50 + "\nSAD PATH") # SAD PATH: 1D example data with torch.inference_mode(): device = "cuda:0" # make example data 1D X = torch.rand((2,), device=device) model = model.to(device) # AOTI compile on CPU example_inputs = X exported_program = torch.export.export(model, (example_inputs,),) exported_model = exported_program.module().to(device) so_path = torch._inductor.aot_compile(exported_model, (example_inputs,),) aot_model = torch._export.aot_load(so_path, device) # Inference: OK print(">> OK: Exported Graph Model: ", exported_model(X)) print(">> OK: AOT Compiled Model: ", aot_model(X)) # Inference: FAILS # NOTE: shape of X_t is invalid X_t = torch.rand((3, 1), device=device) try: exported_model(X_t) except RuntimeError as exc: print(">> FAIL: Exported Graph Model: ", exc) print(">> THIS SHOULD FAIL BUT IS OK: AOT Compiled Model: ", aot_model(X_t)) print("-" * 50) ``` **output** ```bash -------------------------------------------------- HAPPY PATH >> OK: Exported Graph Model: (tensor([[-0.0415]], device='cuda:0'),) >> OK: AOT Compiled Model: tensor([[-0.0415]], device='cuda:0') >> FAIL: Exported Graph Model: Expected input at *args[0].shape[0] to be equal to 1, but got 3 Error: mat1 and mat2 shapes cannot be multiplied (3x1 and 2x1) Exception raised from meta at ../aten/src/ATen/native/LinearAlgebra.cpp:189 (most recent call first): ... >> FAIL: AOT Compiled Model: run_func_( container_handle_, input_handles.data(), input_handles.size(), output_handles.data(), output_handles.size(), cuda_stream_handle, proxy_executor_handle_) API call failed at ../torch/csrc/inductor/aoti_runner/model_container_runner.cpp, line 75 -------------------------------------------------- -------------------------------------------------- SAD PATH >> OK: Exported Graph Model: (tensor([-0.0204], device='cuda:0'),) >> OK: AOT Compiled Model: tensor([-0.0204], device='cuda:0') >> FAIL: Exported Graph Model: Unexpected number of dimensions in input at *args[0].shape (expected torch.Size([2]), got torch.Size([3, 1])) >> THIS SHOULD FAIL BUT IS OK: AOT Compiled Model: tensor([-0.0676], device='cuda:0') -------------------------------------------------- ``` This was observed on CUDA + `torch==2.4.1,2.5.1,nightly`. ### Versions ```bash PyTorch version: 2.4.1+cu121 Is debug build: False CUDA used to build PyTorch: 12.1 ROCM used to build PyTorch: N/A OS: Ubuntu 24.04 LTS (x86_64) GCC version: (Ubuntu 13.2.0-23ubuntu4) 13.2.0 Clang version: Could not collect CMake version: version 3.28.3 Libc version: glibc-2.39 Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0] (64-bit runtime) Python platform: Linux-6.8.0-1018-aws-x86_64-with-glibc2.39 Is CUDA available: True CUDA runtime version: 12.0.140 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA L40S Nvidia driver version: 560.35.03 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 48 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 4 On-line CPU(s) list: 0-3 Vendor ID: AuthenticAMD Model name: AMD EPYC 7R13 Processor CPU family: 25 Model: 1 Thread(s) per core: 2 Core(s) per socket: 2 Socket(s): 1 Stepping: 1 BogoMIPS: 5300.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch topoext ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr rdpru wbnoinvd arat npt nrip_save vaes vpclmulqdq rdpid Hypervisor vendor: KVM Virtualization type: full L1d cache: 64 KiB (2 instances) L1i cache: 64 KiB (2 instances) L2 cache: 1 MiB (2 instances) L3 cache: 8 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-3 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; IBRS_FW; STIBP always-on; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] numpy==2.1.2 [pip3] nvidia-cublas-cu12==12.1.3.1 [pip3] nvidia-cuda-cupti-cu12==12.1.105 [pip3] nvidia-cuda-nvrtc-cu12==12.1.105 [pip3] nvidia-cuda-runtime-cu12==12.1.105 [pip3] nvidia-cudnn-cu12==9.1.0.70 [pip3] nvidia-cufft-cu12==11.0.2.54 [pip3] nvidia-curand-cu12==10.3.2.106 [pip3] nvidia-cusolver-cu12==11.4.5.107 [pip3] nvidia-cusparse-cu12==12.1.0.106 [pip3] nvidia-nccl-cu12==2.20.5 [pip3] nvidia-nvjitlink-cu12==12.6.68 [pip3] nvidia-nvtx-cu12==12.1.105 [pip3] onnxruntime-gpu==1.20.0 [pip3] torch==2.4.1 [pip3] triton==3.0.0 [conda] Could not collect ``` cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4
oncall: pt2,oncall: export
low
Critical
2,674,711,373
godot
handle on exported range property detaches from slider when scrolling while control is hovered
### Tested versions Found in v4.3.stable.official [77dcf97d8] ### System information Godot v4.3.stable - Ubuntu 22.04.5 LTS 22.04 - X11 - Vulkan (Forward+) - dedicated NVIDIA GeForce GTX 970 (nvidia; 550.120) - Intel(R) Core(TM) i9-10900K CPU @ 3.70GHz (20 Threads) ### Issue description This is a visual bug. the slider handles detach from the sliders when hovered while scrolling with the scrollbar https://github.com/user-attachments/assets/b7efe512-a994-4c14-8b30-0212a9dd4737 ### Steps to reproduce 1. go to the inspector on any node or resource with an exported float range property 2. make scrollbar appear and hold click on the scrollbar 3. move mouse over the range slider while scrolling up/down 4. notice the handle detaches ### Minimal reproduction project (MRP) Empty project
bug,topic:gui
low
Critical
2,674,714,085
ollama
OLLAMA_MAX_QUEUE does not limit requests to the same model
### What is the issue? It seems that OLLAMA_MAX_QUEUE is not taking effect. My environment is Windows 11, and I have set OLLAMA_NUM_PARALLEL=1, set OLLAMA_MAX_QUEUE=1, but excessive requests are still queuing up instead of returning an error. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.14
feature request
low
Critical
2,674,734,144
PowerToys
Copy from PowerToys Run error
### Microsoft PowerToys version 0.86.0 ### Installation method WinGet ### Running as admin No ### Area(s) with issue? PowerToys Run ### Steps to reproduce Open PowerToys Run, insert some equation. Click on the "Copy this number to clipboard". Error pops up with following: ![Image](https://github.com/user-attachments/assets/e8d5f9ef-7845-4dc9-843b-fd6920f3af9c) More info in the following bug report summary: [PowerToysReport_2024-11-20-07-38-11.zip](https://github.com/user-attachments/files/17826692/PowerToysReport_2024-11-20-07-38-11.zip) ### โœ”๏ธ Expected Behavior Copy value to clipboard ### โŒ Actual Behavior _No response_ ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Critical
2,674,833,953
pytorch
How to fuse batchnorm to conv2d in the graph exported by torch.export
I used the torch.export to export my CNN model in eval mode,but the op batchnorm still exists. how to eliminate it. Is there some options in torch.export.export function or I should write a fusion pass by myself. Thanks. code: ``` import torch import torch.nn as nn class CNN(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(in_channels=16, out_channels=16, kernel_size=3, stride=1, padding=1) self.bn1 = nn.BatchNorm2d(16) def forward(self, x): x = self.conv1(x) x = self.bn1(x) return x torch.manual_seed(0) model=CNN().eval() input=torch.randn(3,16,224,224) ep=torch.export.export(model,(input,)) print(ep.graph) ``` graph: ``` graph(): %p_conv1_weight : [num_users=1] = placeholder[target=p_conv1_weight] %p_conv1_bias : [num_users=1] = placeholder[target=p_conv1_bias] %p_bn1_weight : [num_users=1] = placeholder[target=p_bn1_weight] %p_bn1_bias : [num_users=1] = placeholder[target=p_bn1_bias] %b_bn1_running_mean : [num_users=1] = placeholder[target=b_bn1_running_mean] %b_bn1_running_var : [num_users=1] = placeholder[target=b_bn1_running_var] %b_bn1_num_batches_tracked : [num_users=0] = placeholder[target=b_bn1_num_batches_tracked] %x : [num_users=1] = placeholder[target=x] %conv2d : [num_users=1] = call_function[target=torch.ops.aten.conv2d.default](args = (%x, %p_conv1_weight, %p_conv1_bias, [1, 1], [1, 1]), kwargs = {}) %_native_batch_norm_legit_no_training : [num_users=1] = call_function[target=torch.ops.aten._native_batch_norm_legit_no_training.default](args = (%conv2d, %p_bn1_weight, %p_bn1_bias, %b_bn1_running_mean, %b_bn1_running_var, 0.1, 1e-05), kwargs = {}) %getitem : [num_users=1] = call_function[target=operator.getitem](args = (%_native_batch_norm_legit_no_training, 0), kwargs = {}) return (getitem,) ``` cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4
oncall: pt2,oncall: export
low
Minor
2,674,880,248
vscode
Settings Editor Synchronization Error
<!-- โš ๏ธโš ๏ธ Do Not Delete This! bug_report_template โš ๏ธโš ๏ธ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- ๐Ÿ•ฎ Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- ๐Ÿ”Ž Search existing issues to avoid creating duplicates. --> <!-- ๐Ÿงช Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- ๐Ÿ’ก Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- ๐Ÿ”ง Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- ๐Ÿช“ If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- ๐Ÿ“ฃ Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.95.3 - OS Version: win 11 Steps to Reproduce: 1. Click the Settings button at the bottom left corner of the Workbench, which opens the Settings Editor window. 2. In the Settings Editor window, modify the visibility of the Workbench's Status Bar. (need) 3. Change the visibility of the status bar through the menu: View > Appearance > Status Bar. 4. At this point, the value of the status bar visibility in the Settings Editor window does not synchronize with the recently set state.
bug,settings-editor,confirmation-pending
low
Critical
2,674,901,219
rust
Compiled wasm32-wasip2 component from simple code requires excessive WASI interfaces
<!-- Thank you for filing a bug report! ๐Ÿ› Please provide a short summary of the bug, along with any information you feel relevant to replicating the bug. --> During my learning and experiments of wasip2, I tried this simple code and compiled it to a wasip2 component: ```rust wit_bindgen::generate!({ // the name of the world in the `*.wit` input file world: "formatter", }); struct Formatter; impl Guest for Formatter { fn format_str(a: String, b: String) -> String { let s = format!("{} + {}", a, b); print(s.as_str()); s } } export!(Formatter); ``` and ``` // format.wit package component:formatter; world formatter { import print: func(s: string); export format-str: func(a: string, b: string) -> string; } ``` The host code, which runs the component, is: ```rust use wasmtime::component::*; use wasmtime::{Engine, Store}; use wasmtime_wasi::{WasiCtx, WasiImpl, WasiView}; use anyhow::Result; // reference: https://docs.rs/wasmtime/latest/wasmtime/component/bindgen_examples/_0_hello_world/index.html // reference: https://docs.wasmtime.dev/examples-rust-wasi.html bindgen!({ path: "../implementation/wit/format.wit", world: "formatter", }); struct MyState { // These two are required basically as a standard way to enable the impl of WasiView wasi_ctx: WasiCtx, table: ResourceTable, } impl WasiView for MyState { fn table(&mut self) -> &mut ResourceTable { &mut self.table } fn ctx(&mut self) -> &mut WasiCtx { &mut self.wasi_ctx } } impl FormatterImports for MyState { fn print(&mut self, s: String) { println!("{}", s); } } /// copied from wasmtime_wasi::type_annotate, which is a private function fn type_annotate<T: WasiView, F>(val: F) -> F where F: Fn(&mut T) -> WasiImpl<&mut T>, { val } fn main() -> Result<()> { let engine = Engine::default(); let component = Component::from_file( &engine, "../implementation/target/wasm32-wasip2/release/implementation.wasm", )?; let mut linker = Linker::new(&engine); let ctx = wasmtime_wasi::WasiCtxBuilder::new().build(); let state = MyState { wasi_ctx: ctx, table: ResourceTable::new(), }; let mut store = Store::new(&engine, state); Formatter::add_to_linker(&mut linker, |s| s)?; // Note: // The below block is copied from `wasmtime_wasi::add_to_linker_sync`. // why a "format!()" in the implementation needs `sync::filesystem::types`, `sync::io::streams`, `cli::exit`, `cli::environment`, `cli::stdin`, `cli::stdout`, `cli::stderr`? { let l = &mut linker; let closure = type_annotate::<MyState, _>(|t| WasiImpl(t)); wasmtime_wasi::bindings::sync::filesystem::types::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::filesystem::preopens::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::io::error::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::sync::io::streams::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::cli::exit::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::cli::environment::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::cli::stdin::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::cli::stdout::add_to_linker_get_host(l, closure)?; wasmtime_wasi::bindings::cli::stderr::add_to_linker_get_host(l, closure)?; } let bindings = Formatter::instantiate(&mut store, &component, &linker)?; let result = bindings.call_format_str(&mut store, "a", "b")?; println!("format_str: {}", result); Ok(()) } ``` The block with a note binds many interfaces to avoid runtime errors that says something like `component imports instance wasi:cli/[email protected], but a matching implementation was not found in the linker.` Removing any one of the lines in the block will result in a runtime error. I expect this compiled component requires none of these WASI interfaces to run, since it has nothing to do with io, cli, etc. Binding these unnecessary interfaces may raise security concerns. The full minimized code is [here](https://github.com/ifsheldon/wit_interface_issue). As a kind person pointed out on [ByteAlliance Zulip](https://bytecodealliance.zulipchat.com/#narrow/channel/217126-wasmtime/topic/Help.3A.20unnecessary.20interface.20requirement.20for.20using.20format!.28.29.3F/near/483417497), these interfaces are required by `std`. Probably there's a way to minimize or prune the interface requirements in the compilation? I think rustc has all the information of which effects are used by any one of functions/macros that is used by a user. At the very lease, I think we should document these requirements somewhere, so there are no hidden/dark interface dependencies that are not specified and unknown in WIT files. ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> `rustc --version --verbose`: ``` rustc 1.82.0 (f6e511eec 2024-10-15) binary: rustc commit-hash: f6e511eec7342f59a25f7c0534f1dbea00d01b14 commit-date: 2024-10-15 host: aarch64-apple-darwin release: 1.82.0 LLVM version: 19.1.1 ```
T-libs,C-discussion,O-wasi
medium
Critical
2,674,935,089
tensorflow
Does tfl.quantize support QI4 data type
**System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 22.04 - TensorFlow installed from (source or binary): source - TensorFlow version (or github SHA if from source): the latest version **Standalone code to reproduce the issue** The [doc of tfl.quantize](https://www.tensorflow.org/mlir/tfl_ops#tflquantize_tflquantizeop) supports QI4 data types, but [the kernel implementation](https://source.chromium.org/chromium/chromium/src/+/main:third_party/tflite/src/tensorflow/lite/kernels/quantize.cc;l=121) doesn't support QI4 data type, the implementation seems not be consistent with the doc.
stat:awaiting tensorflower,type:support,comp:lite
medium
Minor
2,675,055,650
flutter
Linux plugin_test is 2.11% flaky
<!-- meta-tags: To be used by the automation script only, DO NOT MODIFY. { "name": "Linux plugin_test" } --> The post-submit test builder `Linux plugin_test` had a flaky ratio 2.11% for the past (up to) 100 commits, which is above our 2.00% threshold. One recent flaky example for a same commit: https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21536 Commit: https://github.com/flutter/flutter/commit/638331b13328245d7fd5b8cffd729622e3518b8b Flaky builds: https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21536 https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21529 https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21528 Recent test runs: https://flutter-dashboard.appspot.com/#/build?taskFilter=Linux%20plugin_test Please follow https://github.com/flutter/flutter/blob/master/docs/infra/Reducing-Test-Flakiness.md#fixing-flaky-tests to fix the flakiness and enable the test back after validating the fix (internal dashboard to validate: go/flutter_test_flakiness).
platform-android,P1,c: flake,team-android,triaged-android
medium
Major
2,675,235,883
terminal
Releasing both pressed Shift keys (left+right) reports only one
### Windows Terminal version current main ### Windows build number 10.0.19045.5011 ### Other Software _No response_ ### Steps to reproduce 1. Activate "debug tap" pane. - Add the global setting (near the top of your JSON file): ``` "debugFeatures": true ``` - Open a new tab while holding down both left and right Alt . 2. Press and release both Shift keys simultaneously. ### Expected Behavior There are four reports in the "debug tap" pane: - The first Shift key was pressed. - The second Shift key was pressed. - Some Shift key was released. - The remaining Shift key was released. ### Actual Behavior ![Image](https://github.com/user-attachments/assets/e4e03c3d-6d12-4b8e-aa9c-625ffda013c2) There are only three reports in the "debug tap" pane: - The first Shift key was pressed. - The second Shift key was pressed. - Some Shift key was released.
Help Wanted,Area-Input,Issue-Bug,Product-Terminal,Priority-3
low
Critical
2,675,247,756
godot
Can't disable SSL in HttpRequest
### Tested versions v4.4.dev4.mono.official [36e6207bb] ### System information Godot v4.4.dev4.mono - Windows 10.0.26100 - Multi-window, 2 monitors - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2070 (NVIDIA; 32.0.15.6094) - 13th Gen Intel(R) Core(TM) i7-13700 (24 threads) ### Issue description I need post a https. for some reason, it will report TLS_HANDSHAKE_ERROR. i'm test in Godot and Qt. both will report this error. then for Qt, i can add some code for disable tls: ``` QSslConfiguration sslConfig; sslConfig.setPeerVerifyMode(QSslSocket::VerifyNone); objRequest.setSslConfiguration(sslConfig); ``` then Qt can get response successful. but for Godot, i'm try add code like this: ``` http_request.set_tls_options(TLSOptions.client_unsafe()) ``` but unfortunately, it can't work. it always response TLS_HANDSHAKE_ERROR. ### Steps to reproduce none ### Minimal reproduction project (MRP) none ADDTION: i'm search for this issue. but all of them was closed. maybe I'm use this in a wrong way, but i don't know how to do it.
bug,topic:network
low
Critical
2,675,312,557
vscode
Bug Report Must Be Improved
- There are lots of inappropriate issues - Most of them are created via VSCode Issue Reporter screen - The minimum description length is so small - There is no minimum length requirement for Title, 1 character is enough - For example, this is accepted: ![Image](https://github.com/user-attachments/assets/fca0fca5-0f35-467b-ab1e-54896c4bdd13) - Also the issue template can be improved in Github - This is causing many issues be in bad form. For example: https://github.com/microsoft/vscode/issues/234040 https://github.com/microsoft/vscode/issues/234224 https://github.com/microsoft/vscode/issues/234196
feature-request,issue-reporter
low
Critical
2,675,349,542
flutter
ScrollAction slows down significantly when holding down keys
### Steps to reproduce 1. Place a [`MaterialApp`](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget at the root of your widget tree. 2. Place any scrollable widget below the [`MaterialApp`](https://api.flutter.dev/flutter/material/MaterialApp-class.html) and make sure it uses the [`PrimaryScrollController`](https://api.flutter.dev/flutter/widgets/PrimaryScrollController-class.html). 3. Try scrolling down by pressing ctrl+down for desktop or just the down arrow for web. 4. Then try doing the same, but holding the keys down. ### Expected results I would expect to scroll faster when holding down the keys than when pressing them repeatedly. ### Actual results Scrolling becomes significantly slower when holding down the keys compared to pressing them repeatedly. ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() => runApp(const ExampleApp()); class ExampleApp extends StatelessWidget { const ExampleApp({super.key}); @override Widget build(BuildContext context) => MaterialApp( home: Scaffold( body: ListView.builder( primary: true, itemCount: 15, itemBuilder: (BuildContext context, int index) => Container( color: Colors.primaries[index], height: 150, child: Center( child: Text( 'Entry $index', style: Theme.of(context).textTheme.headlineMedium, ), ), ), ), ), ); } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> The first part shows the behavior when pressing the scroll keys repeatedly, and then while holding the scroll keys down. ![ScrollAction Bug Demo](https://github.com/user-attachments/assets/aeb7f07f-a5b6-408f-aa89-394a3c9a2319) </details> ### Logs _No response_ ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [โœ“] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale en_US.UTF-8) โ€ข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter โ€ข Upstream repository https://github.com/flutter/flutter.git โ€ข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800 โ€ข Engine revision a18df97ca5 โ€ข Dart version 3.5.4 โ€ข DevTools version 2.37.3 [โœ—] Android toolchain - develop for Android devices โœ— Unable to locate Android SDK. Install Android Studio from: https://developer.android.com/studio/index.html On first launch it will assist you in installing the Android SDK components. (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). If the Android SDK has been installed to a custom location, please use `flutter config --android-sdk` to update to that location. [โœ“] Chrome - develop for the web โ€ข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable [โœ“] Linux toolchain - develop for Linux desktop โ€ข clang version 19.1.0 (Fedora 19.1.0-1.fc41) โ€ข cmake version 3.30.5 โ€ข ninja version 1.12.1 โ€ข pkg-config version 2.3.0 [!] Android Studio (not installed) โ€ข Android Studio not found; download from https://developer.android.com/studio/index.html (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). [โœ“] VS Code (version 1.95.3) โ€ข VS Code at /usr/share/code โ€ข Flutter extension version 3.100.0 [โœ“] Connected device (2 available) โ€ข Linux (desktop) โ€ข linux โ€ข linux-x64 โ€ข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64 โ€ข Chrome (web) โ€ข chrome โ€ข web-javascript โ€ข Brave Browser 131.1.73.89 [โœ“] Network resources โ€ข All expected network resources are available. ! Doctor found issues in 2 categories. ``` </details>
framework,f: scrolling,platform-web,a: desktop,P2,team-framework,triaged-framework
low
Critical
2,675,480,242
PowerToys
PowerToys RUN crash when I play game
### Microsoft PowerToys version 0.86.0 ### Installation method GitHub ### Running as admin No ### Area(s) with issue? PowerToys Run ### Steps to reproduce I am playing DIRT RALLY 2 with boadless window. Microsoft Windows [10.0.27749.1000] ### โœ”๏ธ Expected Behavior I am starting race. ### โŒ Actual Behavior game screen frozen and PowerToys RUN show error. [2024-11-20.txt](https://github.com/user-attachments/files/17829455/2024-11-20.txt) ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Critical
2,675,501,977
tauri
[feat] How to modify the referer attribute of video and audio on the Android platform
### Describe the problem How to modify the referer attribute of video and audio on the Android platform, because I am now making the app version of the website, but in Tauri, the referer looks like tarui.local ### Describe the solution you'd like How to modify the referer attribute of video and audio on the Android platform, because I am now making the app version of the website, but in Tauri, the referer looks like tarui.local ### Alternatives considered _No response_ ### Additional context _No response_
type: feature request
low
Minor
2,675,502,131
flutter
NestedScrollView's body is not scrollable using keyboard shortcuts with the ScrollAction
### Steps to reproduce 1. Place a [MaterialApp](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget at the root of your widget tree. 2. Place a [NestedScrollView](https://api.flutter.dev/flutter/widgets/NestedScrollView-class.html) below the [MaterialApp](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget. 3. Try scrolling down by pressing ctrl+down for desktop or just the down arrow for web. ### Expected results I would expect the header and then the body of the nested scroll view to be scrolled down. ### Actual results The header is scrolled down as expected, but when reaching its extent, the body is not scrolled. ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() => runApp(const ExampleApp()); class ExampleApp extends StatelessWidget { const ExampleApp({super.key}); @override Widget build(BuildContext context) => MaterialApp( home: Scaffold( body: NestedScrollView( headerSliverBuilder: ( BuildContext context, bool innerBoxIsScrolled, ) => <Widget>[ SliverOverlapAbsorber( handle: NestedScrollView.sliverOverlapAbsorberHandleFor( context, ), sliver: SliverAppBar( pinned: true, expandedHeight: 150.0, forceElevated: innerBoxIsScrolled, flexibleSpace: FlexibleSpaceBar( title: const Text('SliverAppBar'), background: Container(color: Colors.red), centerTitle: true, ), ), ), ], body: CustomScrollView( slivers: <Widget>[ Builder( builder: (context) => SliverOverlapInjector( handle: NestedScrollView.sliverOverlapAbsorberHandleFor( context, ), ), ), SliverFixedExtentList( itemExtent: 48.0, delegate: SliverChildBuilderDelegate( (BuildContext context, int index) => ListTile( title: Text('Item $index'), ), childCount: 30, ), ), ], ), ), ), ); } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> The following video shows the scrolling behavior when using a trackpad. ![Scrolling a NestedScrollView using a trackpad](https://github.com/user-attachments/assets/94924f51-e1ee-4268-a76c-644e0a953a33) We'd expect similar scrolling behavior when using the arrow keys, but instead the scrolling stops when the scroll extent of the header is reached. ![Scrolling a NestedScrollView using keyboard shortcuts](https://github.com/user-attachments/assets/799ca701-8096-4569-afc5-b5f9f612a49f) </details> ### Logs _No response_ ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [โœ“] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale en_US.UTF-8) โ€ข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter โ€ข Upstream repository https://github.com/flutter/flutter.git โ€ข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800 โ€ข Engine revision a18df97ca5 โ€ข Dart version 3.5.4 โ€ข DevTools version 2.37.3 [โœ—] Android toolchain - develop for Android devices โœ— Unable to locate Android SDK. Install Android Studio from: https://developer.android.com/studio/index.html On first launch it will assist you in installing the Android SDK components. (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). If the Android SDK has been installed to a custom location, please use `flutter config --android-sdk` to update to that location. [โœ“] Chrome - develop for the web โ€ข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable [โœ“] Linux toolchain - develop for Linux desktop โ€ข clang version 19.1.0 (Fedora 19.1.0-1.fc41) โ€ข cmake version 3.30.5 โ€ข ninja version 1.12.1 โ€ข pkg-config version 2.3.0 [!] Android Studio (not installed) โ€ข Android Studio not found; download from https://developer.android.com/studio/index.html (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). [โœ“] VS Code (version 1.95.3) โ€ข VS Code at /usr/share/code โ€ข Flutter extension version 3.100.0 [โœ“] Connected device (2 available) โ€ข Linux (desktop) โ€ข linux โ€ข linux-x64 โ€ข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64 โ€ข Chrome (web) โ€ข chrome โ€ข web-javascript โ€ข Brave Browser 131.1.73.89 [โœ“] Network resources โ€ข All expected network resources are available. ! Doctor found issues in 2 categories. ``` </details>
framework,f: scrolling,platform-web,a: desktop,f: focus,has reproducible steps,P3,team-framework,triaged-framework,found in release: 3.24,found in release: 3.27
low
Minor
2,675,504,350
ui
[bug]: Can't install shadcn
### Describe the bug npx shadcn@latest init -d >> C:\Users\Hs\Documents\sam> โœ” Preflight checks. โœ” Verifying framework. Found Next.js. โœ” Validating Tailwind CSS. โœ” Validating import alias. โœ” Writing components.json. โœ” Checking registry. โœ” Updating tailwind.config.js โœ” Updating src\app\globals.css โ ‹ Installing dependencies. Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge spawn UNKNOWN ### Affected component/components the component itself ### How to reproduce npx shadcn@latest init -d >> C:\Users\Hs\Documents\sam> โœ” Preflight checks. โœ” Verifying framework. Found Next.js. โœ” Validating Tailwind CSS. โœ” Validating import alias. โœ” Writing components.json. โœ” Checking registry. โœ” Updating tailwind.config.js โœ” Updating src\app\globals.css โ ‹ Installing dependencies. Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge spawn UNKNOWN ### Codesandbox/StackBlitz link _No response_ ### Logs ```bash npx shadcn@latest init -d >> C:\Users\Hs\Documents\sam> โœ” Preflight checks. โœ” Verifying framework. Found Next.js. โœ” Validating Tailwind CSS. โœ” Validating import alias. โœ” Writing components.json. โœ” Checking registry. โœ” Updating tailwind.config.js โœ” Updating src\app\globals.css โ ‹ Installing dependencies. Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge spawn UNKNOWN ``` ### System Info ```bash npx shadcn@latest init -d >> C:\Users\Hs\Documents\sam> โœ” Preflight checks. โœ” Verifying framework. Found Next.js. โœ” Validating Tailwind CSS. โœ” Validating import alias. โœ” Writing components.json. โœ” Checking registry. โœ” Updating tailwind.config.js โœ” Updating src\app\globals.css โ ‹ Installing dependencies. Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge spawn UNKNOWN ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,675,508,933
deno
Deno.serve timeouts on AWS lambda
Version: Deno 2.0.6 I'm trying to deploy a **complied** deno project into AWS Lambda. I'm using Hono, and in the docs, they are using [Deno.serve](https://hono.dev/docs/getting-started/deno#_3-hello-world). The problem is that lambda timeouts and returns `Internal Server Error`. After investigating the problem it seems that the `Deno.serve` is causing this. In the log I can see: ``` 2024-11-20T09:22:13.960Z | Listening on http://0.0.0.0:8000/ 2024-11-20T09:22:22.838Z | EXTENSION Name: lambda-adapter State: Ready Events: [] 2024-11-20T09:22:22.838Z | INIT_REPORT Init Duration: 9999.92 ms Phase: init Status: timeout ``` Steps: 1. Create a `main.ts` with `Deno.serve((req) => new Response("Hello World!"));` 2. Build & deploy the app: ``` deno compile -A --target x86_64-unknown-linux-gnu --no-check -o bootstrap main.ts && chmod +x ./bootstrap && zip -r -D artifact.zip bootstrap && aws lambda update-function-code --region eu-central-1 --function-name deno-compiled --zip-file fileb://artifact.zip ``` 3. Remember to add lambda layer adapter `arn:aws:lambda:eu-central-1:753240598075:layer:LambdaAdapterLayerX86:23`. I know there is a dockerised approach for deploying Deno, but I don't want to use it because it uses too much memory. If you take an oak example from [aws-lambda-web-adapter examples](https://github.com/awslabs/aws-lambda-web-adapter/blob/main/examples/deno-zip/src/main.ts), compile it, and deploy it to AWS Lambda - it works.
needs investigation
low
Critical
2,675,606,238
TypeScript
NoInfer not working on callback parameter destructuring
### ๐Ÿ”Ž Search Terms NoInfer parameter callback destructuring ### ๐Ÿ•— Version & Regression Information Latest to date. ### โฏ Playground Link https://www.typescriptlang.org/play/?#code/MYGwhgzhAEBiD28A8AVANNASgPmgbwCgBIYeAOwgBcAnAV2EvmoApqBTMAE3JAE9pgAIwBc0ZpVHpoAfVEA5eAEkyAMzbUkeaAA9RZWgFtB6gNzR4lABbr5S1eqQ5oAX2wBKaAF5cmD3mfEAPSBAuRUdAxMrBzcZHwCImIS0FKy+Dp6hsbUZhbW1LbKahpOzh7eWH4BRMGhFDT0jCzsXDz8QqLikhhpCkUOWrrQ+kam5lY2WC7uXj5VBAEEpPUyAIxew2wA7nCIzMxkYAZsouEAlmQA5hiDLuW4zFqHxxjad25mtQjI51cYtGQANZkeBbMjYAi1CCWeC0ECcaDGXY-GgXa7pZ4naC-S5mIYjbLTAhLMKUGQAJg2ZG2yP2mNOqKu9zETyObHenxCWyYgIgom+SBxN2GbIZ1DRRKAA ### ๐Ÿ’ป Code ```ts class Foo<T, R> { constructor(readonly cb: (t: T, _: NoInfer<{ x: number; other: NoInfer<R> }>) => R) {} } const _1 = new Foo((name: string, { x }) => ({ name, x })); // Foo<string, unknown> // should be Foo<string, { name: string; x: number }> const _2 = new Foo((name: string) => ({ name })); // works: Foo<string, { name: string }> ``` ### ๐Ÿ™ Actual behavior infers `unknown`. ### ๐Ÿ™‚ Expected behavior Should ignore the parameter tagged with `NoInfer` and infer correctly. ### Additional information about the issue Is this intended behavior? feels like a bug to me. If the callback parameter is destructured partially it infers `unknown`, but it works if the parameter is omitted. I tried `NoInfer` at various locations.
Help Wanted,Possible Improvement
medium
Critical
2,675,669,443
pytorch
[Dynamo] Dynamo is not aware of the function users override
### ๐Ÿ› Describe the bug Hi, we are trying to enabling Llama in Intel HPU, but found graph break in `torch.cuda.current_device()` . We have a package GPU migration tool, it replaces `torch.cuda.current_device()` to `torch.hpu.current_device()`. But we found dynamo is not aware of the function override, and still consider `torch.cuda.current_device()`, eventually leading to graph break or program failed. Can you give us some suggestions or have any plan to fix it? ### Error logs Here is a simple case ```python import os os.environ["PT_HPU_LAZY_MODE"] = "0" import torch import habana_frameworks.torch.dynamo.compile_backend def hpu_device(): return torch.hpu.current_device() torch.cuda.current_device = hpu_device @torch.compile(backend="hpu_backend", fullgraph=True) def foo(x): dim_size = list(x.size()) y = torch.tensor([6], device=torch.cuda.current_device()) return torch.mul(x, y) x = torch.tensor([6], device=torch.hpu.current_device()) foo(x) ``` Our GPU migration tool implement, override cuda current_device(): ```python @CudaModuleRegister.register_f() def current_device() -> int: """ .. py:gpumgrcall:: current_device.hpu_match Maps torch.cuda.current_device to torch.hpu.current_device. """ G_LOGGER.info( api_type="hpu_match", func_prefix="torch.cuda", new_call="torch.hpu.current_device()", ) return torch.hpu.current_device() ``` The error gets from dynamo: ``` File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/bytecode_transformation.py", line 1322, in transform_code_object transformations(instructions, code_options) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 219, in _fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 634, in transform tracer.run() File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 2796, in run super().run() File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 983, in run while self.step(): File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 895, in step self.dispatch_table[inst.opcode](self, inst) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 582, in wrapper return inner_fn(self, inst) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1602, in CALL_FUNCTION self.call_function(fn, args, {}) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 830, in call_function self.push(fn.call_function(self, args, kwargs)) # type: ignore[arg-type] File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/functions.py", line 727, in call_function unimplemented(msg) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/exc.py", line 297, in unimplemented raise Unsupported(msg, case_name=case_name) torch._dynamo.exc.Unsupported: 'skip function hpu_device in file /root/test.py' from user code: File "/root/test.py", line 14, in foo y = torch.tensor([6], device=torch.cuda.current_device()) ``` ### Versions ``` PyTorch version: 2.5.1a0+git3ba5690 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-5.15.0-122-generic-x86_64-with-glibc2.35 Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 160 On-line CPU(s) list: 0-159 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Platinum 8380 CPU @ 2.30GHz CPU family: 6 Model: 106 Thread(s) per core: 2 Core(s) per socket: 40 Socket(s): 2 Versions of relevant libraries: [pip3] habana-torch-dataloader==1.19.0.428 [pip3] habana-torch-plugin==1.19.0.428 [pip3] numpy==1.23.5 [pip3] pytorch-lightning==2.4.0 [pip3] torch==2.5.1a0+git3ba5690 [pip3] torch_tb_profiler==0.4.0 [pip3] torchaudio==2.5.1a0+1661daf [pip3] torchdata==0.9.0+d4bb3e6 [pip3] torchmetrics==1.6.0 [pip3] torchtext==0.18.0a0+9bed85d [pip3] torchvision==0.20.1a0+3ac97aa [conda] Could not collect ``` cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @amjames
triaged,oncall: pt2,module: dynamo
low
Critical
2,675,693,600
flutter
Unable to scroll diagonally with a trackpad in a TwoDimensionalScrollView
### Steps to reproduce 1. Build the example application from the ["2D scrolling - Flutter Build Show"](https://www.youtube.com/watch?v=ppEdTo-VGcg). 2. Try scrolling diagonally using a trackpad. ### Expected results I would expect the viewport to change offset in both axes simultaneously. ### Actual results The viewport changes its offset in a single axis only. ### Code sample <details open><summary>Code sample</summary> [DartPad - Two Dimensional Grid in Flutter](https://dartpad.dev/?id=4424936c57ed13093eb389123383e894) You can additionaly wrap the scroll view in a [Listener](https://api.flutter.dev/flutter/widgets/Listener-class.html) to make sure the scroll events have offset in both axes. ```dart ... body: Listener( onPointerSignal: (PointerSignalEvent signal) { if (signal is PointerScrollEvent) { print(signal.scrollDelta); } }, child: TwoDimensionalGridView( ... ``` Output sample: ```console flutter: Offset(-7.8, -14.4) flutter: Offset(-6.1, -14.4) flutter: Offset(-7.0, -14.4) flutter: Offset(-6.1, -14.4) ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> Diagonal drag scrolling works as expected: ![Scrolling a 2D scroll view diagonally using drag gestures](https://github.com/user-attachments/assets/69e0c7dd-2916-414d-bdfe-c61edd37bad1) Diagonal pointer scrolls are not applied on both axes at the same time: ![Scrolling a 2D scroll view diagonally using a trackpad](https://github.com/user-attachments/assets/8bca8ceb-ce69-41e2-bcbb-c71be5dfd320) </details> ### Logs _No response_ ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [โœ“] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale en_US.UTF-8) โ€ข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter โ€ข Upstream repository https://github.com/flutter/flutter.git โ€ข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800 โ€ข Engine revision a18df97ca5 โ€ข Dart version 3.5.4 โ€ข DevTools version 2.37.3 [โœ—] Android toolchain - develop for Android devices โœ— Unable to locate Android SDK. Install Android Studio from: https://developer.android.com/studio/index.html On first launch it will assist you in installing the Android SDK components. (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). If the Android SDK has been installed to a custom location, please use `flutter config --android-sdk` to update to that location. [โœ“] Chrome - develop for the web โ€ข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable [โœ“] Linux toolchain - develop for Linux desktop โ€ข clang version 19.1.0 (Fedora 19.1.0-1.fc41) โ€ข cmake version 3.30.5 โ€ข ninja version 1.12.1 โ€ข pkg-config version 2.3.0 [!] Android Studio (not installed) โ€ข Android Studio not found; download from https://developer.android.com/studio/index.html (or visit https://flutter.dev/to/linux-android-setup for detailed instructions). [โœ“] VS Code (version 1.95.3) โ€ข VS Code at /usr/share/code โ€ข Flutter extension version 3.100.0 [โœ“] Connected device (2 available) โ€ข Linux (desktop) โ€ข linux โ€ข linux-x64 โ€ข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64 โ€ข Chrome (web) โ€ข chrome โ€ข web-javascript โ€ข Brave Browser 131.1.73.89 [โœ“] Network resources โ€ข All expected network resources are available. ! Doctor found issues in 2 categories. ``` </details>
framework,f: scrolling,platform-web,a: desktop,P2,team-framework,triaged-framework
low
Minor
2,675,698,485
vscode
Presentation of experimental settings in settings.json
![Image](https://github.com/user-attachments/assets/d2662f4c-f1d9-4542-a770-8958e4b38e06) Thoughts: 1. Why the line about no quick fixes? 2. I was expecting it to look more like deprecated, where there's a yellow underline to be much more obvious about settings that could potentially break the editor.
bug,settings-editor
low
Minor
2,675,717,701
vscode
Light bulb shows up for move to new file?
Maybe I remember wrong but didn't we hide the lightbulb when the only actions were move to file? ![Image](https://github.com/user-attachments/assets/62412826-b28d-4847-92fb-1e4edfdbac40) Version: 1.96.0-insider (user setup) Commit: 8d1d1373cd9b8408222b51958e9688827a9e27ce Date: 2024-11-20T09:23:45.992Z Electron: 32.2.5 ElectronBuildId: 10579404 Chromium: 128.0.6613.186 Node.js: 20.18.0 V8: 12.8.374.38-electron.0 OS: Windows_NT x64 10.0.26100
bug,editor-code-actions
low
Minor
2,675,759,556
kubernetes
Cronjobs occasionally run much later than scheduled
### What happened? CronJobs will occasionally fail to create their Jobs at the scheduled time, with delays ranging from 3 minutes to over 50 minutes. In the most extreme cases we've observed delays over 60 minutes. From the kube-controller-manager logs, if I'm understanding correctly (which i may not), it appears that CronJobs rely exclusively on the reconciliation process to trigger Job creation. If no reconciliation occurs around the scheduled time, the Job is not created until a subsequent reconciliation cycle is triggered. Logs consistently show the following pattern (verbosity set to 8), even for CronJobs that execute on time, this job ran 5 minutes later than scheduled: ```console I1120 11:05:32.562672 1 cronjob_controllerv2.go:526] "No unmet start times" logger="cronjob-controller" cronjob="camel-test/mkt-cloud-segments-servicecloud-v2-simu-buidar-pp" I1120 11:05:32.562702 1 cronjob_controllerv2.go:219] "Re-queuing cronjob" logger="cronjob-controller" cronjob="camel-test/mkt-cloud-segments-servicecloud-v2-simu-buidar-pp" requeueAfter="23h54m27.538988861s" ``` Ocasionally this pattern can also be observed in the kube-controller-manager logs: ```console E1120 13:33:05.831607 1 cronjob_controllerv2.go:168] error syncing CronJobController golang/datafeeds-download-datafeeds, requeuing: Operation cannot be fulfilled on cronjobs.batch "datafeeds-download-datafeeds": the object has been modified; please apply your changes to the latest version and try again ``` Performing manual actions, such as editing the CronJob object or deleting an old Job, forces an immediate reconciliation and triggers the creation of the pending Job. Another thing we observed is that the issue only affects cronjobs that run on a daily basis (these are distributed at different times) but those that run frequently, like every 15 minutes, never get delayed. Things we checked: 1. Control-plane clocks verified: All clocks are synchronized using NTP. No drift observed. 2. Increased logging verbosity: Enabled v=8 for kube-controller-manager 3. The number of cronjobs does not seem to be too much for the size of the cluster and resource utilization We don't know what could be causing this behavior and we don't know how to continue debugging it further. ### What did you expect to happen? CronJobs should create their corresponding Jobs at the scheduled time. ### How can we reproduce it (as minimally and precisely as possible)? These are the details of the basic current cluster configuration, however I do not know if the issue could be replicated in another similar cluster. - Cluster Size: 8 worker nodes, 3 control-plane nodes - CronJob API Version: batch/v1 - Number of CronJobs: ~100 scheduled at various intervals, distributed in multiple namespaces Resource Utilization: - Control-plane nodes have sufficient CPU and memory resources (not near limits). - The kube-controller-manager pod does not have resource limits set. ### Anything else we need to know? Keda operator is installed and we have 3 deployments that are managed by Keda, logs regarding these objects are frequent: ```console 11:46:17.755947 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=0 metric="external metric s0-cron-Europe-Madrid-07xx1-5-021xx1-5(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-activemq-dev,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:17 +0100 CET" scaleTarget="Deployment/activemq-test/activemq-dev" 11:46:17.760043 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=1 metric="external metric s0-cron-Europe-Madrid-0000xxx-5523xxx(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-integracions-erpex,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:17 +0100 CET" scaleTarget="Deployment/camel/integracions-erpex" 11:46:31.385874 graph_builder.go:681] "GraphBuilder process object" logger="garbage-collector-controller" apiVersion="keda.sh/v1alpha1" kind="ScaledObject" object="camel/cron-integracions-erpex" uid="d9d7f22a-0eef-4c21-8c23-7823fc6e1f60" eventType="update" virtual=false 11:46:31.900165 graph_builder.go:681] "GraphBuilder process object" logger="garbage-collector-controller" apiVersion="keda.sh/v1alpha1" kind="ScaledObject" object="activemq-test/cron-activemq-dev" uid="eef2ed20-c9ef-40a2-bee4-3c0fcc71dd5a" eventType="update" virtual=false 11:46:32.782034 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=0 metric="external metric s0-cron-Europe-Madrid-07xx1-5-021xx1-5(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-activemq-dev,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:32 +0100 CET" scaleTarget="Deployment/activemq-test/activemq-dev" 11:46:32.787861 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=1 metric="external metric s0-cron-Europe-Madrid-0000xxx-5523xxx(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-integracions-erpex,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:32 +0100 CET" scaleTarget="Deployment/camel/integracions-erpex" ``` There is a custom operator that observes events on several custom resources. We've observed this issue through multiple kubernetes versions, no upgrade has fixed it. ### Kubernetes version <details> ```console $ kubectl version Client Version: v1.30.6 Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3 Server Version: v1.30.6 ``` </details> ### Cloud provider <details> On premise </details> ### OS version <details> ```console # On Linux: PRETTY_NAME="Ubuntu 22.04.5 LTS" NAME="Ubuntu" VERSION_ID="22.04" VERSION="22.04.5 LTS (Jammy Jellyfish)" VERSION_CODENAME=jammy ID=ubuntu ID_LIKE=debian HOME_URL="https://www.ubuntu.com/" SUPPORT_URL="https://help.ubuntu.com/" BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" UBUNTU_CODENAME=jammy $ uname -a Linux k8scp02 5.15.0-119-generic #129-Ubuntu SMP Fri Aug 2 19:25:20 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux ``` </details> ### Install tools <details> kubeadm </details> ### Container runtime (CRI) and version (if applicable) <details> containerd 1.7.12-0ubuntu2~22.04.1 </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> vwmare-csi v3.3.1 cilium v1.16.1 </details>
kind/bug,sig/apps,needs-triage
low
Critical
2,675,801,028
go
os/signal: TestSignalTrace is flaky on openbsd/arm64
The `os/signal.TestSignalTrace` is flaky on openbsd/arm64: https://build.golang.org/log/5a334b71cd825842917d59f362ca9c211da82ec9 https://build.golang.org/log/2c4c9b74d2a13ab71e0a00d8430fbf367a4a96e1 This appears to be a signal being lost - it's not currently clear when this issue started, or where the problem exists (and it may well be a kernel related issue). The failure can be reproduced by running `os/signal.TestSignalTrace` in a loop on an Apple M1 or Apple M2. Running another competing workload (such as a build) will speed up the test failure. This test should be skipped until it can be diagnosed further.
OS-OpenBSD,NeedsInvestigation,arch-arm64,compiler/runtime
low
Critical
2,675,889,804
ui
[bug]: component at https://ui.shadcn.com/r/colors/Gray.json was not found.
### Describe the bug The component at https://ui.shadcn.com/r/colors/Gray.json was not found. It may not exist at the registry. Please make sure it is a valid component. ### Affected component/components Drawer ### How to reproduce 1. Go to Docs on Shadcn 2. Click on Drawer 3. follow the installation steps ### Codesandbox/StackBlitz link _No response_ ### Logs _No response_ ### System Info ```bash Terminal ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,675,916,345
angular
adev: Dropdown design issue on Firefox
### Describe the problem that you experienced Not sure if all dropdowns are affected, but the dropdown on the [Angular Update page](https://angular.dev/update-guide) has two issues on Firefox: - There is some horizontal scrolling (which starts all the way right, making it impossible to read what it says without scrolling left). - The scrollbar background is translucent. ### Enter the URL of the topic with the problem https://angular.dev/update-guide ### Describe what you were looking for in the documentation _No response_ ### Describe the actions that led you to experience the problem _No response_ ### Describe what you want to experience that would fix the problem _No response_ ### Add a screenshot if that helps illustrate the problem Firefox: ![Image](https://github.com/user-attachments/assets/2edb27f7-7652-41f4-b939-dcb194039e8b) Chrome: ![Image](https://github.com/user-attachments/assets/e132b3fe-7ed0-4925-81a3-5ee5f9a5174b) ### If this problem caused an exception or error, please paste it here ```true ``` ### If the problem is browser-specific, please specify the device, OS, browser, and version ```true Firefox 132 (on Windows 10, not sure if relevant) ``` ### Provide any additional information here in as much as detail as you can ```true ```
area: docs-infra
low
Critical
2,675,939,472
pytorch
[FlexAttention] Block sparse mask is significantly slower than sdpa and handmade triton implementation
### ๐Ÿ› Describe the bug When testing flex attention with the mask from issue #141126, I found that apart from being incorrect it is also significantly slower than torch's generic `scaled_dot_product_attention` in the example (at lower seq lens < 1000). ![streaming-attention-context-32-back-10](https://github.com/user-attachments/assets/20566f83-308d-4443-8892-bce7ea09de64) Also, it's way slower than handmade (and correct) triton implementation of the same attention. Ask me if any additional info is needed. ### Versions PyTorch version: 2.5.0 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.4 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.11.10 (main, Sep 7 2024, 18:35:41) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-5.4.210-39.1-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: Could not collect CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA A800-SXM4-80GB GPU 1: NVIDIA A800-SXM4-80GB GPU 2: NVIDIA A800-SXM4-80GB GPU 3: NVIDIA A800-SXM4-80GB GPU 4: NVIDIA A800-SXM4-80GB GPU 5: NVIDIA A800-SXM4-80GB GPU 6: NVIDIA A800-SXM4-80GB GPU 7: NVIDIA A800-SXM4-80GB Nvidia driver version: 550.54.14 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.1.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.1.0 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: False CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 48 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 120 On-line CPU(s) list: 0-119 Vendor ID: AuthenticAMD Model name: AMD EPYC 7662 64-Core Processor CPU family: 23 Model: 49 Thread(s) per core: 1 Core(s) per socket: 1 Socket(s): 120 Stepping: 0 BogoMIPS: 3992.18 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities Virtualization: AMD-V L1d cache: 7.5 MiB (120 instances) L1i cache: 7.5 MiB (120 instances) L2 cache: 60 MiB (120 instances) L3 cache: 1.9 GiB (120 instances) NUMA node(s): 4 NUMA node0 CPU(s): 0-29 NUMA node1 CPU(s): 30-59 NUMA node2 CPU(s): 60-89 NUMA node3 CPU(s): 90-119 Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP disabled, RSB filling, PBRSB-eIBRS Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.24.3 [pip3] torch==2.5.0 [pip3] torch-tb-profiler==0.4.3 [pip3] triton==3.1.0 [conda] Could not collect cc @msaroufim @jbschlosser @bhosmer @cpuhrsch @erichan1 @drisspg @mikaylagawarecki @Chillee @yanboliang @BoyuanFeng @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225
module: performance,triaged,module: flex attention,module: sdpa
low
Critical
2,675,957,062
material-ui
[docs] Convert the "all components" images to CSS
### Summary The [Material UI](https://mui.com/material-ui/all-components/), [Base UI](https://mui.com/base-ui/all-components/), and [Charts](https://mui.com/x/react-charts/#all-mui-x-charts-components) docs have an "all components" page with illustrated cards for each component. These illustrations are currently done on Figma and added to the cards as png. That's not entirely ideal because: 1. Maintaining them on Figma is an additional tool to manage 2. Compressing images and then needing to download them is cumbersome and not efficient performance-wise 3. We can't add micro animations So, the idea is to, at some point, tackle all of these points by converting them into CSS. This issue tracks the effort for the Base UI images. **Search keywords**: component images css **Search keywords**:
docs,design
low
Major
2,675,961,578
flutter
Build linux failed by INSTALL not found "native_assets/linux"
### Steps to reproduce 'flutter build linux' will trigger this bug. I've tried 'flutter clean' or create a sample project but this bug still occurred. ### Expected results build successfully just as 'flutter build apk' does. ### Actual results ``` CMake Error at cmake_install.cmake:112 (file): file INSTALL cannot find "/home/....../build/native_assets/linux": No such file or directory. ``` ### Code sample <details open><summary>Code sample</summary> ```dart [Paste your code here] ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] </details> ### Logs <details open><summary>Logs</summary> ```console [Paste your logs here] ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console Doctor summary (to see all details, run flutter doctor -v): [โœ“] Flutter (Channel stable, 3.24.5, on Arch Linux 6.11.9-zen1-1-zen, locale en_US.UTF-8) [!] Android toolchain - develop for Android devices (Android SDK version 34.0.0) ! Some Android licenses not accepted. To resolve this, run: flutter doctor --android-licenses [โœ“] Chrome - develop for the web [โœ“] Linux toolchain - develop for Linux desktop [!] Android Studio (not installed) [โœ“] IntelliJ IDEA Ultimate Edition (version 2024.3) [โœ“] Connected device (2 available) [โœ“] Network resources ``` </details>
platform-linux,a: desktop,P2,team-linux,triaged-linux
low
Critical
2,675,966,815
material-ui
[Autocomplete][docs] The Tailwind code demo doesn't include item focus styles
### Related page https://mui.com/base-ui/react-autocomplete/ ### Kind of issue Missing information ### Issue description The Tailwind autocomplete demo has visible focus on selected items, i.e. if you hover an item or used the up/down keys, items are visibly focused. However, if you copy the code from the code sample into a project, or if you open the [linked codesandbox](https://codesandbox.io/s/myqs2l?file=/src/Demo.tsx), those visible focus states don't work and there's no visible feedback about which item is focused. From the devtools, it looks like there are some extra css rules applied to the in-page demo that aren't in the code sample. ### Context I'm trying to use the base-ui autocomplete component with Tailwind in the way demonstrated on the page. **Search keywords**: tailwind focus autocomplete base-ui **Search keywords**:
bug ๐Ÿ›,docs,component: autocomplete
low
Major
2,675,978,447
ant-design
[BUG] Select with search requires double enter hit to trigger the onChange event
### Reproduction link [![Edit on StackBlitz](https://developer.stackblitz.com/img/open_in_stackblitz.svg)](https://stackblitz.com/edit/antd-reproduce-5x-qe3zcq?file=demo.tsx) ### Steps to reproduce 1. Choose 'Edoardo' user shown into the Select by click 2. Re-open the Select and fill the text with 'Brugola', then hit 'Enter' (one hit the search works) 3. Re-open the Select and fill the text with 'Pistoni', then hit 'Enter' (not work at the first hit. Hit 'Enter' again and it works) ### What is expected? Hit the 'Enter' button only once time. ### What is actually happening? After the first search by text (that works as expected), the Select requires two hit by 'Enter' button to perform the onChange event. | Environment | Info | | --- | --- | | antd | 5.22.1 | | React | 18.3.1 | | System | nodeJS latest | | Browser | Chrome latest | --- This is happened with React state and REDUX also. <!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
๐Ÿ› Bug,Inactive
low
Critical
2,675,981,770
vscode
Horizontal scroll position resets in file when seek peek definition called
<!-- โš ๏ธโš ๏ธ Do Not Delete This! bug_report_template โš ๏ธโš ๏ธ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- ๐Ÿ•ฎ Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- ๐Ÿ”Ž Search existing issues to avoid creating duplicates. --> <!-- ๐Ÿงช Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- ๐Ÿ’ก Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- ๐Ÿ”ง Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- ๐Ÿช“ If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- ๐Ÿ“ฃ Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.95.2 - OS Version: linux mint mate 21.3 Version: 1.95.2 Commit: e8653663e8840adaf45af01eab5c627a5af81807 Date: 2024-11-07T11:07:22.054Z Electron: 32.2.1 ElectronBuildId: 10427718 Chromium: 128.0.6613.186 Node.js: 20.18.0 V8: 12.8.374.38-electron.0 OS: Linux x64 5.15.0-126-generic __Issue:__ With long lines in a code file and horizontally scrolled to bring the remaining part of the line into view In this scroll position, when tried to seek peek definition the seek peek definition shows inline, but the original code file horizontally scrolls back to start position Steps to Reproduce: 1. Identify long code line in a file and horizontally move to bring a symbol to view 2. select seek peek definition for this symbol sample images - ![Image](https://github.com/user-attachments/assets/2de853df-32de-4585-b0f1-f1d17b704b00) ![Image](https://github.com/user-attachments/assets/41d014a1-9bb8-4351-ab4d-5c2c9e1e8e56) ( note the horizontal scroll position has come to start position ) and have scroll back to see symbol in code. __Expected behaviour:__ horizontal scroll position to be retained during seek peek definition as in sample image below ![Image](https://github.com/user-attachments/assets/fb1a4243-2c09-4203-a031-03acfc608e38)
feature-request,ux,polish,editor-symbols
low
Critical
2,675,982,410
rustdesk
Keyboard German to German Raspberry PI4 not working
### Bug Description Hello, If i connect with Rustdesk to a Raspberry PI4 local Language German: ![image](https://github.com/user-attachments/assets/829a2e66-7f81-4932-a27d-4a0138f4e0d2) The Keys are not forwarded correctly. For example instead of * i get }. Do you have a Idea how to fix this? Thanks ### How to Reproduce Install Rustdesk under Raspbian and connect. ### Expected Behavior Correct keys. ### Operating system(s) on local (controlling) side and remote (controlled) side Windows 10 -> Raspbian ### RustDesk Version(s) on local (controlling) side and remote (controlled) side 1.3.1 - > 1.2.7 ### Screenshots ![image](https://github.com/user-attachments/assets/6cafb1af-a945-4be2-b9b8-258ed8f57995) ### Additional Context _No response_
bug
low
Critical
2,675,996,407
vscode
Enable highlighting scroll in copilot chat window
<!-- โš ๏ธโš ๏ธ Do Not Delete This! feature_request_template โš ๏ธโš ๏ธ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> In the Github Copilot chat window it can sometimes give very long answers, rewriting entire files/scripts, when only a portion of that code is actually changing. Currently, when I try to highlight a longer section of code but don't want all of it, I need to click the start of where I want to copy, scroll down manually, and click the end with shift held to highlight everything. Instead it would be easier if I could highlight and have the scrollbar move with the highlighted section as it hits the top/bottom of the currently visible window. This type of feature is implemented in most other programs with scroll text boxes.
chat
low
Minor
2,676,046,424
node
FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal
### Version v22.11.0 ### Platform ```text Darwin 23.6.0 Darwin Kernel Version 23.6.0 I also tested it on Linux machine. ``` ### Subsystem _No response_ ### What steps will reproduce the bug? Hi, I want to report a node bug. It can reproduced by running the following PoC, node.js gives a FATAL error. ```javascript const {spawn} = require('child_process'); Object.defineProperty(Array.prototype, "0", { set(){ console.log(123); } }); spawn("ls"); ``` Please note that this bug differs from the reported [here](https://github.com/nodejs/node/issues/54186). AH. ### How often does it reproduce? Is there a required condition? By just running the given code, node runtime gives node crash. ### What is the expected behavior? Why is that the expected behavior? Not a crash, by looking at the stack trace, it is connected to `v8::ToLocalChecked Empty MaybeLocal`. ### What do you see instead? FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal ----- Native stack trace ----- 1: 0x10bdd2277 node::OnFatalError(char const*, char const*) 2: 0x10c025171 v8::Utils::ReportApiFailure(char const*, char const*) 3: 0x10bef84cc node::(anonymous namespace)::ProcessWrap::Spawn(v8::FunctionCallbackInfo<v8::Value> const&) 4: 0x10cb21a62 Builtins_CallApiCallbackGeneric [~/node/v22.11.0/bin/node] ----- JavaScript stack trace ----- 1: ChildProcess.spawn (node:internal/child_process:395:28) 2: spawn (node:child_process:759:9) 3: ${PATH_TO_POC} 4: node:internal/modules/cjs/loader:1546:14 5: node:internal/modules/cjs/loader:1689:10 6: node:internal/modules/cjs/loader:1318:32 7: node:internal/modules/cjs/loader:1128:12 8: traceSync (node:diagnostics_channel:315:14) 9: wrapModuleLoad (node:internal/modules/cjs/loader:218:24) 10: executeUserEntryPoint (node:internal/modules/run_main:170:5) ### Additional information _No response_
confirmed-bug,child_process
low
Critical
2,676,131,157
ollama
High Inference Time and Limited GPU Utilization with Ollama Docker
### What is the issue? ## Description: I am using Ollama in a Docker setup with GPU support, configured to use all available GPUs on my system. However, when using the NemoTron model with a simple prompt and utilizing the function calling feature, the inference time is around 50 seconds to get a response, which is too high for my use case. ## Docker Configuration: Here is my docker-compose.yml file for Ollama: services: ollama: image: ollama/ollama container_name: ollama hostname: ollama ports: - "11434:11434" volumes: - /home/system/dockers/volumes/ollama:/root/.ollama runtime: nvidia environment: - NVIDIA_VISIBLE_DEVICES=all deploy: resources: reservations: devices: - driver: nvidia count: all capabilities: [gpu] restart: unless-stopped networks: - genai_network networks: genai_network: ## Log Details: ### Server Initialization: Listening on [::]:11434 (version 0.4.2) Dynamic LLM libraries runners="[cpu_avx2 cuda_v11 cuda_v12 cpu cpu_avx]" Looking for compatible GPUs Inference compute id=GPU-660ca8b7-181a-ede9-f6fe-8ccd5f9dbb89 library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="47.1 GiB" Inference compute id=GPU-d62f5e11-4192-0e70-0732-55b558edcb7a library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="46.4 GiB" Inference compute id=GPU-45972459-815c-f304-9fdf-b952276c9b13 library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="47.0 GiB" ### Model Loading: ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA RTX 6000 Ada Generation, compute capability 8.9, VMM: yes llm_load_tensors: ggml ctx size = 0.68 MiB ## Observed Issue: During inference with the NemoTron model, the response time is around 50 seconds. The logs show that only one GPU seems to be utilized (found 1 CUDA devices), despite multiple GPUs being detected during service initialization. ## Questions: ### GPU Load Balancing: Does Ollama support load balancing across multiple GPUs? If yes, why do the logs indicate that only one GPU is used (found 1 CUDA devices) when the model is loaded? ### Performance Optimization: What steps are recommended to reduce inference time? Should I adjust configuration settings in Docker or Ollama? Could variables like GGML_CUDA_FORCE_CUBLAS or GGML_CUDA_FORCE_MMQ improve performance? ## Technical Context: Ollama Version: 0.4.2 Hardware Configuration: GPUs: 3 x NVIDIA RTX 6000 Ada Generation (47.5 GiB VRAM each) NVIDIA Driver: 12.4 Model Used: NemoTron Usage Scenario: Simple prompt with function calling. ## Expectation: - Confirmation on multi-GPU support in Ollama. - Suggestions to reduce inference time. - Documentation or examples of optimized configuration for heavy workloads with multiple GPUs. ## Thank You for Your Help! ![nvidia-smi](https://github.com/user-attachments/assets/b3be6b97-0b16-47ed-8942-2ff7f4dd6aff) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.4.2
bug
low
Major
2,676,135,808
flutter
[go_router] Browser back button unexpectedly navigates to an already replaced route that was the result of a redirection
### Steps to reproduce 1. Launch the sample 2. Tap the "Go to child screen" button 3. Wait 2 seconds for the deep link screen to forward you to the child screen 4. Being on the child screen, tap the browser's back button ### Expected results You see the home screen ### Actual results You see the deep link screen ### Code sample Link to Dartpad https://dartpad.dev/?id=72e8664b6408534f2448cc8a34199ee6 <details open><summary>Code sample</summary> ```dart import 'package:flutter/foundation.dart'; import 'package:flutter/material.dart'; import 'package:go_router/go_router.dart'; // 1. The main file void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); @override Widget build(BuildContext context) { return MaterialApp.router(routerConfig: goRouter); } } // 2. The GoRouter configuration final goRouter = GoRouter( debugLogDiagnostics: kDebugMode, routes: [ GoRoute( path: '/', builder: (context, state) => const SplashScreen(), ), GoRoute( path: '/home', builder: (context, state) => const HomeScreen(), ), GoRoute( path: '/deeplink', builder: (context, state) => const DeepLinkScreen(), ), GoRoute( path: '/child', builder: (context, state) => const ChildScreen(), redirect: (context, GoRouterState state) { final valueAddedDuringDeepLink = state.uri.queryParameters['value']; if (valueAddedDuringDeepLink == '42') { return null; } else { return '/deeplink'; } }, ), ], ); // 3. The 4 screens // 3.1 Splash screen class SplashScreen extends StatefulWidget { const SplashScreen({super.key}); @override State<SplashScreen> createState() => _SplashScreenState(); } class _SplashScreenState extends State<SplashScreen> { @override void initState() { super.initState(); Future.delayed(const Duration(seconds: 1), () { context.go('/home'); }); } @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar(title: const Text('Splash screen')), body: const Center( child: Column( mainAxisSize: MainAxisSize.min, children: [ Text('Loading...'), CircularProgressIndicator(), ], )), ); } } // 3.2 Home screen class HomeScreen extends StatelessWidget { const HomeScreen({super.key}); @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar(title: const Text('Home screen')), body: Center( child: Column( mainAxisSize: MainAxisSize.min, children: [ const Text('Home'), TextButton( onPressed: () => context.push('/child'), child: const Text('Go to child screen'), ), ], )), ); } } // 3.3 Deep link screen class DeepLinkScreen extends StatefulWidget { const DeepLinkScreen({super.key}); @override State<DeepLinkScreen> createState() => _DeepLinkScreenState(); } class _DeepLinkScreenState extends State<DeepLinkScreen> { @override void initState() { super.initState(); Future.delayed(const Duration(seconds: 1), () { context.pushReplacement('/child?value=42'); }); } @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar(title: const Text('Deep link screen')), body: const Center( child: Column( mainAxisSize: MainAxisSize.min, children: [ Text('Redirecting to the child screen...'), CircularProgressIndicator(), ], ))); } } // 3.4 Child screen class ChildScreen extends StatelessWidget { const ChildScreen({super.key}); @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar(title: const Text('Child screen')), body: const Center(child: Text('Child')), ); } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> https://github.com/user-attachments/assets/fcfe8c3e-bbe6-4747-b864-0039a2e858c9 </details> ### Logs <details open><summary>Logs</summary> ```console Launching lib\main.dart on Chrome in debug mode... Waiting for connection from debug service on Chrome... This app is linked to the debug service: ws://127.0.0.1:63040/Dkr73jPlCtM=/ws Debug service listening on ws://127.0.0.1:63040/Dkr73jPlCtM=/ws Debug service listening on ws://127.0.0.1:63040/Dkr73jPlCtM=/ws [GoRouter] Full paths for routes: โ”œโ”€/ (SplashScreen) โ”œโ”€/home (HomeScreen) โ”œโ”€/deeplink (DeepLinkScreen) โ””โ”€/child (ChildScreen) [GoRouter] setting initial location null [GoRouter] Using MaterialApp configuration [GoRouter] going to /home [GoRouter] pushing /child [GoRouter] redirecting to RouteMatchList#7876f(uri: /deeplink, matches: [RouteMatch#f0590(route: GoRoute#6038c(name: null, path: "/deeplink"))]) [GoRouter] pushReplacement /child?value=42 [GoRouter] pushing /child [GoRouter] redirecting to RouteMatchList#7876f(uri: /deeplink, matches: [RouteMatch#f0590(route: GoRoute#6038c(name: null, path: "/deeplink"))]) [GoRouter] pushReplacement /child?value=42 [GoRouter] pushReplacement /child?value=42 ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console PS G:\Projects\gorouter-playground> flutter doctor -v [โˆš] Flutter (Channel stable, 3.19.4, on Microsoft Windows [Version 10.0.22631.4460], locale en-US) โ€ข Flutter version 3.19.4 on channel stable at C:\Development\SDKs\Flutter โ€ข Upstream repository https://github.com/flutter/flutter.git โ€ข Framework revision 68bfaea224 (8 months ago), 2024-03-20 15:36:31 -0700 โ€ข Engine revision a5c24f538d โ€ข Dart version 3.3.2 โ€ข DevTools version 2.31.1 [โˆš] Windows Version (Installed version of Windows is version 10 or higher) [โˆš] Android toolchain - develop for Android devices (Android SDK version 34.0.0) โ€ข Android SDK at C:\Development\SDKs\Android โ€ข Platform android-34, build-tools 34.0.0 โ€ข ANDROID_HOME = C:\Development\SDKs\Android โ€ข ANDROID_SDK_ROOT = C:\Development\SDKs\Android โ€ข Java binary at: C:\Development\IDEs\AndroidStudioKoala\jbr\bin\java โ€ข Java version OpenJDK Runtime Environment (build 17.0.11+0--11852314) โ€ข All Android licenses accepted. [โˆš] Chrome - develop for the web โ€ข Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe [โˆš] Visual Studio - develop Windows apps (Visual Studio Community 2022 17.11.6) โ€ข Visual Studio at C:\Program Files\Microsoft Visual Studio\2022\Community โ€ข Visual Studio Community 2022 version 17.11.35431.28 โ€ข Windows 10 SDK version 10.0.22621.0 [โˆš] Android Studio (version 2024.1) โ€ข Android Studio at C:\Development\IDEs\AndroidStudioKoala โ€ข Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter โ€ข Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart โ€ข Java version OpenJDK Runtime Environment (build 17.0.11+0--11852314) [โˆš] IntelliJ IDEA Community Edition (version 2022.1) โ€ข IntelliJ at C:\Development\IDEs\IntelliJ IDEA\IntelliJ IDEA Community Edition 2020.3.2 โ€ข Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter โ€ข Dart plugin version 221.6103.1 [โˆš] Connected device (3 available) โ€ข Windows (desktop) โ€ข windows โ€ข windows-x64 โ€ข Microsoft Windows [Version 10.0.22631.4460] โ€ข Chrome (web) โ€ข chrome โ€ข web-javascript โ€ข Google Chrome 131.0.6778.70 โ€ข Edge (web) โ€ข edge โ€ข web-javascript โ€ข Microsoft Edge 130.0.2849.68 [โˆš] Network resources โ€ข All expected network resources are available. โ€ข No issues found! ``` </details>
platform-web,package,has reproducible steps,P2,p: go_router,team-go_router,triaged-go_router,found in release: 3.24,found in release: 3.27
low
Critical
2,676,138,749
TypeScript
MIDIMessageEvent data is incorrectly typed
### ๐Ÿ”Ž Search Terms In https://github.com/microsoft/TypeScript/pull/57027 `MIDIMessageEvent.data` was changed from `Uint8Array` to `Uint8Array | null` (see https://github.com/microsoft/TypeScript/pull/57027/files#diff-dc0eab937d15e62545da3ed7b4f40ad6b24f15dd88fbc6ceda2bfb4ed8356eb0L14670). According to https://developer.mozilla.org/en-US/docs/Web/API/MIDIMessageEvent/data data the value will always be an Uint8Array. ### ๐Ÿ•— Version & Regression Information This changed in #57027 ### โฏ Playground Link _No response_ ### ๐Ÿ’ป Code ### ๐Ÿ™ Actual behavior `'event.data' is possibly 'null'.ts(18047)` ### ๐Ÿ™‚ Expected behavior No warning about data property possibly being null. ### Additional information about the issue _No response_
Bug,Help Wanted,Domain: lib.d.ts
low
Minor
2,676,203,778
react
[React 19] useOptimistic - Add option to disable rollback
## Summary `useOptimistic` automatically rolls back upon error. Can we make this optional? Use case: I would like to keep the optimistic data displaying upon error, and display a "Try again" button next to it. (yes, I could do this without useOptimistic, but I suspect this will be a common use case). Also, the current rollback on error behavior [isn't documented here](https://react.dev/reference/react/useOptimistic), so suggest adding before release. <!-- Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a repository on GitHub, or provide a minimal code example that reproduces the problem. You may provide a screenshot of the application if you think it is relevant to your bug report. Here are some tips for providing a minimal example: https://stackoverflow.com/help/mcve. -->
React 19
medium
Critical
2,676,240,652
godot
Performance drop in 3D scenes with DirectionalLight in Godot 3.6 compared to 3.5
### Tested versions - Godot 3.5.3 - Godot 3.6 ### System information Linux Ubuntu 22.04, Intel Core 2 Duo 2.8 Ghz, Geforce 9500 GT 3Gb VRAM, 4 Gb RAM 800 Mhz ### Issue description Performance is considerably worse in Godot 3.6, compared to 3.5 Godot 3.5.3, with 33 FPS: ![Image](https://github.com/user-attachments/assets/ccd9f534-edcf-4dea-bfc7-8df4e8986413) Same exact scene in Godot 3.6 with 27 FPS: ![Image](https://github.com/user-attachments/assets/208af5b8-22c9-4a17-bb06-24b79ca2bf14) A drop of 20% in performance can be observed. ### Steps to reproduce Run a project in Godot 3.5.3, make note of the FPS Run the same project in Godot 3.6 and observe the drop in performance. ### Minimal reproduction project (MRP) The truck town demo can be used to attest the performance change. https://github.com/godotengine/godot-demo-projects/releases/download/3.5-9e68af3/3d_truck_town.zip
bug,topic:rendering,regression,topic:3d,performance
low
Major
2,676,249,674
ollama
Ollama on AMD CPU iGPU Windows Hack
Followed this tutorial https://github.com/ollama/ollama/pull/5059 and got Ollama running on windows on AMD CPU, any tips on how I get it to use iGPU what commands can i try to hack? Thanks
feature request
low
Minor
2,676,290,606
rust
Tracking Issue for `vec_deque_iter_as_slices`
<!-- Thank you for creating a tracking issue! Tracking issues are for tracking a feature from implementation to stabilization. Make sure to include the relevant RFC for the feature if it has one. If the new feature is small, it may be fine to skip the RFC process. In that case, you can use `issue = "none"` in your initial implementation PR. The reviewer will ask you to open a tracking issue if they agree your feature can be added without an RFC. --> Feature gate: `#![feature(vec_deque_iter_as_slices)]` This is a tracking issue for: - `alloc::collections::vec_deque::Iter::as_slices` - `alloc::collections::vec_deque::IterMut::into_slices` - `alloc::collections::vec_deque::IterMut::as_slices` - `alloc::collections::vec_deque::IterMut::as_mut_slices` <!-- Include a short description of the feature. --> Each of these functions obtains a pair of slices from a `VecDeque` iterator, similar to `VecDeque::as_slices`. ### Public API <!-- For most library features, it'd be useful to include a summarized version of the public API. (E.g. just the public function signatures without their doc comments or implementation.) --> ```rust use std::collections::vec_deque::{Iter, IterMut}; impl<'a, T> Iter<'a, T> { pub fn as_slices(&self) -> (&'a [T], &'a [T]) { ... } } impl<'a, T> IterMut<'a, T> { pub fn into_slices(self) -> (&'a mut [T], &'a mut [T]) { ... } pub fn as_slices(&self) -> (&[T], &[T]) { ... } pub fn as_mut_slices(&mut self) -> (&mut [T], &mut [T]) { ... } } ``` ### Steps / History <!-- For larger features, more steps might be involved. If the feature is changed later, please add those PRs here as well. --> - [x] Implementation: #123947 - [ ] Final comment period (FCP)[^1] - [ ] Stabilization PR <!-- Once the feature has gone through a few release cycles and there are no unresolved questions left, the feature might be ready for stabilization. If this feature didn't go through the RFC process, a final comment period (FCP) is always needed before stabilization. This works as follows: A library API team member can kick off the stabilization process, at which point the rfcbot will ask all the team members to verify they agree with stabilization. Once enough members agree and there are no concerns, the final comment period begins: this issue will be marked as such and will be listed in the next This Week in Rust newsletter. If no blocking concerns are raised in that period of 10 days, a stabilization PR can be opened by anyone. --> ### Unresolved Questions <!-- Include any open questions that need to be answered before the feature can be stabilised. If multiple (unrelated) big questions come up, it can be a good idea to open a separate issue for each, to make it easier to keep track of the discussions. It's useful to link any relevant discussions and conclusions (whether on GitHub, Zulip, or the internals forum) here. --> - None yet. [^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html
T-libs-api,C-tracking-issue
low
Minor
2,676,331,380
excalidraw
Increase text element padding in rectangle
As of now, the text padding inside the rectangle is way too small, so it's often necessary to manually add the spaces to achieve a more pleasant look. Instead, we shall increase the padding, so that no manual "alignment" is not necessary. Things to keep in mind: - it might be a regression -> double-check - do not create a breaking change for existing diagrams (after a manual redraw, like entering wysiwyg, it's likely fine) - don't forget about padding while editing (wysiwyg) - test different sizes & zoom levels ![Image](https://github.com/user-attachments/assets/e98c94a9-a8da-4a0f-ae3c-b3a8347d93de)
good first issue,UX/UI
low
Major
2,676,337,729
flutter
[web] [Pointer Interceptor] App crashed on Chrome when closing a dialog containing Pointer Interceptor and Voice over is turn on
### Steps to reproduce 1. Run the app 2. Turn Voice over on 3. Click to open the dialog 4. Close the dialog from the Enable or Disabled button ### Expected results Navigated back to initial screen ### Actual results Crash the app on Chrome with Error code 11 ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; import 'package:pointer_interceptor/pointer_interceptor.dart'; Future<void> main() async => runApp(const ShowDialogExampleApp()); class ShowDialogExampleApp extends StatelessWidget { const ShowDialogExampleApp({super.key}); @override Widget build(BuildContext context) { return const MaterialApp( home: DialogExample(), ); } } class DialogExample extends StatelessWidget { const DialogExample({super.key}); @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar(title: const Text('showDialog Sample')), body: Center( child: OutlinedButton( onPressed: () => _dialogBuilder(context), child: const Text('Open Dialog'), ), ), ); } Future<void> _dialogBuilder(BuildContext context) { return showDialog<void>( context: context, builder: (BuildContext context) { return AlertDialog( title: const Text('Basic dialog title'), content: const Text( 'A dialog is a type of modal window that\n' 'appears in front of app content to\n' 'provide critical information, or prompt\n' 'for a decision to be made.', ), actions: <Widget>[ PointerInterceptor( child: TextButton( style: TextButton.styleFrom( textStyle: Theme.of(context).textTheme.labelLarge, ), child: const Text('Disable'), onPressed: () { Navigator.of(context).pop(); }, ), ), TextButton( style: TextButton.styleFrom( textStyle: Theme.of(context).textTheme.labelLarge, ), child: const Text('Enable'), onPressed: () { Navigator.of(context).pop(); }, ), ], ); }, ); } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] </details> <img width="1840" alt="Screenshot 2024-11-20 at 17 12 26" src="https://github.com/user-attachments/assets/71b54cd7-5816-49d0-880a-202fc3f8126c"> <img width="1840" alt="Screenshot 2024-11-20 at 17 13 11" src="https://github.com/user-attachments/assets/fbe3ce7a-f5e9-4423-a7f0-9481af2c631f"> <img width="1728" alt="Screenshot 2024-11-20 at 17 13 48" src="https://github.com/user-attachments/assets/c483da20-eea3-492f-b00c-3b1bcf1eb9fe"> ### Logs <details open><summary>Logs</summary> ```console Debug service listening on ws://127.0.0.1:62102/PNkeCMZovkk=/ws Debug service listening on ws://127.0.0.1:62102/PNkeCMZovkk=/ws Debugger: Target crashed! Debugger: Target crashed! Debugger: Target crashed! ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [Paste your output here] ``` </details>
c: crash,a: accessibility,platform-web,package,has reproducible steps,P2,p: pointer_interceptor,team-web,triaged-web,found in release: 3.27,found in release: 3.28
low
Critical
2,676,379,804
tensorflow
The documentation in `data_performance.ipynb` uses `py_function()` without an explanation
### Issue type Documentation Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version TF 2.18 ### Custom code No ### OS platform and distribution _No response_ ### Mobile device _No response_ ### Python version _No response_ ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? In the ["Better performance with the tf.data API" guide](https://www.tensorflow.org/guide/data_performance), `tf.py_function()` is used several times in the mapping function. There are various performance boosts demonstrated by the guide with said mapping functions. However, there does not seem to be any practical reason why `tf.py_function()` is used inside the mapping functions. In fact, if you remove those, the behavior is the same; in other words, there doesn't seem to be a need for them at all. Curiously, if you remove them from the examples and then perform the time measurements, the speedup goes away. For example, consider the following mapping function from [the guide](https://github.com/tensorflow/docs/blob/bbc0b9c70fc0bd4411793d1b0bcc56ef1dbc2405/site/en/guide/data_performance.ipynb#L447-L450): ```python def mapped_function(s): # Do some hard pre-processing tf.py_function(lambda: time.sleep(0.03), [], ()) return s ``` In the sequential case, it is shown that the following mapping code results in a time measurement of 0.49222556600034295: ```python benchmark( ArtificialDataset() .map(mapped_function) ) ``` And, the optimized (parallel) version as follows results in a time measurement of 0.36392719900049997: ```python benchmark( ArtificialDataset() .map( mapped_function, num_parallel_calls=tf.data.AUTOTUNE ) ) ``` But, if I remove the `tf.py_function()` from the mapping function, I get comparable measurements from both examples, namely, 0.22448736599994845 and 0.2266392660001202: ```python def mapped_function(s): # Do some hard pre-processing lambda: time.sleep(0.03), [], () return s ``` In fact, that measurement is even better, which makes me believe that this example is contrived to show a performance benefit by using `num_parallel_calls` when in fact TF is already optimizing the code without it. Frivolously wrapping the function in `tf.py_function()` is most likely causing TensorFlow *not* to optimize the function. Thus, is `num_parallel_calls` even needed to achieve better performance? ### Standalone code to reproduce the issue ```shell Mentioned in the above description. ``` ### Relevant log output _No response_
type:docs-bug,type:bug
low
Critical
2,676,399,871
terminal
Rendering with tcell under WSL is slow for the first 65s
### Windows Terminal version 1.23.3231.0 ### Windows build number 10.0.26100.0 ### Other Software _No response_ ### Steps to reproduce Under WSL run this: https://github.com/gdamore/tcell/blob/main/_demos/stress.go ### Expected Behavior _No response_ ### Actual Behavior Initially the application runs super slow. After 65 seconds it suddenly becomes extremely fast (= the expected performance). Here you can see the transition between the two: https://github.com/user-attachments/assets/765f5cbe-88b2-4c7e-a21c-bbdefb555687
Product-Conpty,Issue-Bug,Product-Terminal,Needs-Tag-Fix,Priority-2
low
Major
2,676,443,002
kubernetes
Adopting new libraries from cel-go
### What would you like to be added? CEL is general used in Kubernetes now. We wanna keep up with new libraries added into cel-go. cel recently released the [new release](https://github.com/google/cel-go/releases/tag/v0.22.0) which has some feature we love to include in Kubernetes. This issue is to track the adoption of [email protected]. More info: cel introduced versioning for newly added libraries. Kubernetes would love to keep up with supporting the libraries in versions. ### Why is this needed? /sig api-machinery /triage accepted
sig/api-machinery,kind/feature,triage/accepted
low
Minor
2,676,445,140
pytorch
CuDNN SDPA Issue Tracker
# Summary Keeps track of known issues and the current state of the integration, with the goal to bump order in priority list with higher confidence ```[tasklist] ### Tasks - [ ] https://github.com/pytorch/pytorch/issues/139298 - [x] Match Query's stride layout https://github.com/pytorch/pytorch/pull/138354 - [x] gradOutput vs Ouput stride mismatch, blocked by cudnn bump: https://github.com/pytorch/pytorch/pull/138354#issuecomment-2488863424 - [ ] Fill Value twiddiling: https://github.com/pytorch/pytorch/pull/140837 - [ ] Debugging PR: https://github.com/pytorch/pytorch/pull/139699 - [ ] Update CuDNN backend to respect enable_gqa https://github.com/pytorch/pytorch/issues/139586 - [ ] Mitigated: https://github.com/pytorch/pytorch/issues/138529 but should see if fixed in new version of cudnn - [ ] WSL: https://github.com/huggingface/diffusers/issues/9704 | need to fix or guard correctly ``` cc @csarofeen @ptrblck @xwang233 @eqy @msaroufim @mikaylagawarecki
module: cudnn,module: cuda,triaged,module: sdpa
low
Critical
2,676,447,125
PowerToys
Time Picker for Awake - 12H option
### Description of the new feature / enhancement Can we add option for 24H vs 12H option for picking end time of awake expire? ### Scenario when this would be used? Choosing end time for Awake expire on specific time and date. ### Supporting information _No response_
Idea-Enhancement,Needs-Triage,Product-Awake
low
Minor
2,676,456,476
go
testing: panic in AllocsPerRun during parallel test
I propose that testing.AllocsPerRun panic if called during a parallel test, because it cannot give accurate results in that context. Tests that use the result in that context will be flaky. For example, flake #70327 was caused by a test that called t.Parallel and then used AllocsPerRun. The other tests running in parallel had allocations that were observed by the AllocsPerRun, resulting in inconsistent results and a flaky failure. I've implemented this change in [CL 630137](https://go.dev/cl/630137).
Proposal,Proposal-Accepted
low
Critical
2,676,482,668
kubernetes
Migrating cel ext library cost close to the func
### What would you like to be added? Current we keep the extension libraries added in Kubernetes located at: https://github.com/kubernetes/kubernetes/tree/master/staging/src/k8s.io/apiserver/pkg/cel/library All the cost defined for the ext library is defined in https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apiserver/pkg/cel/library/cost.go which is separately from the definition of func. Future contributors who is not familiar with the ext library would easily missing adding cost while adding new library. It would be great to have cost defined closer to the func. ### Why is this needed? Good for code maintenance. /sig api-machinery /triage accepted
sig/api-machinery,kind/feature,triage/accepted
low
Minor
2,676,595,627
godot
Error when using Move/Duplicate To
### Tested versions Engine version: Godot Engine v4.4.dev4.official (36e6207bb48fafd43b98a458d0c461ad982b4e68) ### System information Godot v4.4.dev4 - Pop!_OS 22.04 LTS on X11 - X11 display driver, Multi-window, 2 monitors - Vulkan (Forward+) - dedicated Quadro P1000 (nvidia; 560.35.03) - Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz (12 threads) ### Issue description Godot crashes when trying to copy a file to another location using the Move'/Duplicate To option in the FileSystem. This is the error I get: ``` ERROR: Task '_update_scan_actions' already exists. at: add_task (editor/progress_dialog.cpp:173) ================================================================ handle_crash: Program crashed with signal 11 Engine version: Godot Engine v4.4.dev4.official (36e6207bb48fafd43b98a458d0c461ad982b4e68) Dumping the backtrace. Please include this when reporting the bug to the project developer. [1] /lib/x86_64-linux-gnu/libc.so.6(+0x42520) [0x7c0f70e42520] (??:0) [2] /usr/local/bin/godot() [0x1588dd0] (??:0) [3] /usr/local/bin/godot() [0x158b9c3] (??:0) [4] /usr/local/bin/godot() [0x4215a28] (??:0) [5] /usr/local/bin/godot() [0x25ec59c] (??:0) [6] /usr/local/bin/godot() [0x2602c93] (??:0) [7] /usr/local/bin/godot() [0x2606159] (??:0) [8] /usr/local/bin/godot() [0x52ee92] (??:0) [9] /usr/local/bin/godot() [0x184fe53] (??:0) [10] /usr/local/bin/godot() [0x16677c1] (??:0) [11] /usr/local/bin/godot() [0x1589215] (??:0) [12] /usr/local/bin/godot() [0x158bf78] (??:0) [13] /usr/local/bin/godot() [0x4264c7d] (??:0) [14] /usr/local/bin/godot() [0x26060fc] (??:0) [15] /usr/local/bin/godot() [0x52ee92] (??:0) [16] /usr/local/bin/godot() [0x41c2fa] (??:0) [17] /lib/x86_64-linux-gnu/libc.so.6(+0x29d90) [0x7c0f70e29d90] (??:0) [18] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0x80) [0x7c0f70e29e40] (??:0) [19] /usr/local/bin/godot() [0x43af1a] (??:0) -- END OF BACKTRACE -- ================================================================ ``` ### Steps to reproduce Right click in one file in the FileSystem section and chose Move/Duplicate To option. Then select another location in the dialog and chose Copy on the options below. Godot crashes completely. ### Minimal reproduction project (MRP) I'm using an empty 2D project with just a few folders and files, nothing special.
bug,topic:editor,needs testing,crash
low
Critical
2,676,609,698
bitcoin
Functional tests: `feature_bind_port_discover.py` is failing
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current behaviour The test is failing due to `Discover()` not being able to finding the local addresses "1.1.1.1" and "2.2.2.2". ### Expected behaviour The test should pass. Addr1 and Addr2 should be found. ### Steps to reproduce Run the following commands: ```bash sudo ifconfig lo:0 1.1.1.1/32 up && sudo ifconfig lo:1 2.2.2.2/32 up ./build/test/functional/test_runner.py --ihave1111and2222 feature_bind_port_discover.py ``` ### Relevant log output ```bash andre@acer:~/repos/bitcoin$ build/test/functional/test_runner.py --ihave1111and2222 feature_bind_port_discover.py Temporary test directory at /tmp/test_runner_โ‚ฟ_๐Ÿƒ_20241120_141406 Remaining jobs: [feature_bind_port_discover.py] 1/1 - feature_bind_port_discover.py failed, Duration: 1 s stdout: 2024-11-20T17:14:06.192000Z TestFramework (INFO): PRNG seed is: 4142310399432710311 2024-11-20T17:14:06.193000Z TestFramework (INFO): Initializing test directory /tmp/test_runner_โ‚ฟ_๐Ÿƒ_20241120_141406/feature_bind_port_discover_0 2024-11-20T17:14:06.511000Z TestFramework (INFO): Test that if -bind= is not passed then all addresses are added to localaddresses 2024-11-20T17:14:06.512000Z TestFramework (ERROR): Assertion failed Traceback (most recent call last): File "/home/andre/repos/bitcoin/test/functional/test_framework/test_framework.py", line 132, in main self.run_test() File "/home/andre/repos/bitcoin/build/test/functional/feature_bind_port_discover.py", line 63, in run_test assert found_addr1 AssertionError 2024-11-20T17:14:06.563000Z TestFramework (INFO): Stopping nodes 2024-11-20T17:14:06.716000Z TestFramework (WARNING): Not cleaning up dir /tmp/test_runner_โ‚ฟ_๐Ÿƒ_20241120_141406/feature_bind_port_discover_0 2024-11-20T17:14:06.716000Z TestFramework (ERROR): Test failed. Test logging available at /tmp/test_runner_โ‚ฟ_๐Ÿƒ_20241120_141406/feature_bind_port_discover_0/test_framework.log 2024-11-20T17:14:06.716000Z TestFramework (ERROR): 2024-11-20T17:14:06.716000Z TestFramework (ERROR): Hint: Call /home/andre/repos/bitcoin/test/functional/combine_logs.py '/tmp/test_runner_โ‚ฟ_๐Ÿƒ_20241120_141406/feature_bind_port_discover_0' to consolidate all logs 2024-11-20T17:14:06.716000Z TestFramework (ERROR): 2024-11-20T17:14:06.716000Z TestFramework (ERROR): If this failure happened unexpectedly or intermittently, please file a bug and provide a link or upload of the combined log. 2024-11-20T17:14:06.716000Z TestFramework (ERROR): https://github.com/bitcoin/bitcoin/issues 2024-11-20T17:14:06.716000Z TestFramework (ERROR): stderr: TEST | STATUS | DURATION feature_bind_port_discover.py | โœ– Failed | 1 s ALL | โœ– Failed | 1 s (accumulated) Runtime: 1 s ``` ### How did you obtain Bitcoin Core Compiled from source ### What version of Bitcoin Core are you using? Bitcoin Core daemon version v28.99.0-22ef95dbe3e4 ### Operating system and version Ubuntu 22.04 ### Machine specifications _No response_
Tests,CI failed
low
Critical
2,676,610,602
flutter
[Android] Consider implementing backwards compatibility for sensitive content protection
Flutter is currently working to support sensitive context protection on Android (see https://github.com/flutter/flutter/issues/150218) which is based on the [View#setContentSensitivity](https://developer.android.com/reference/android/view/View#setContentSensitivity(int)) API that is only supported on Android 15+. We should consider implementing some form of backwards compatibility for lower API versions by trying to detect when the screen is being shared by using the [DisplayManager](https://developer.android.com/reference/android/hardware/display/DisplayManager) to detect active `Display`s. Note that this would also require figuring out how to obscure sensitive content if we are able to detect screen sharing on Android APIs < 15. See comment on technical limitations section in go/flutter-android-sensitive-content for more context.
c: new feature,platform-android,engine,P3,team-android,triaged-android
low
Minor
2,676,620,744
pytorch
Consider noting the PyTorch version for BC-breaking error messages
I hit the following error message in torch.load https://github.com/pytorch/pytorch/blob/8d708090c0eb306facfd8f85d58c578a8cbbe689/torch/serialization.py#L1338-L1341 while working on enabling a model with torch.compile for user empathy day. It took me a couple of minutes to parse the error message and to determine if it was my fault, the code's fault, or PyTorch's fault. My conclusion was that this code used to succeed in the past but we have changed it to fail in a recent release (or on main). I would have appreciated an error message to that effect, like: "In PyTorch 2.x we changed the behavior of torch.load to only load weights. This torch.load failed because it serialized more than just weights. If you're confident that the file is safe to load, then you may pass weights_only=False to torch.load)." cc @ezyang @gchanan @mikaylagawarecki
module: bc-breaking,triaged,topic: improvements
low
Critical
2,676,625,500
tauri
[bug] BSOD rarely when exiting the application
### Describe the bug Hello My application causes a BSOD, error code โ€œVideo_scheduler_internal_errorโ€, very rarely (about once out of 70 times) when it is exited, either by code, by alt+F4, by reloading the code or by clicking on the exit button The event observer also gives the indication โ€œThe driver detected an internal driver error on \Device\VBoxNetLwf.โ€ My dependencies are : serde_json = โ€œ1.0 serde = { version = โ€œ1.0โ€, features = [โ€œderiveโ€] } tokio = { version = โ€œ1โ€, features = [โ€œfullโ€] } tauri = { version = โ€œ1.8.1โ€, features = [โ€œapi-allโ€] } reqwest = { version = โ€œ0.11โ€, features = [โ€œjsonโ€, โ€œrustls-tlsโ€] } futures = โ€œ0.3โ€ zip = โ€œ2.1.6โ€ My application is developed on Windows 10, npm and javascript. Version of tauri-cli 2.1.0 Let me know if you need any further information ### Reproduction Quitting the application gives it a small but real chance of causing a BSOD; it doesn't matter how you quit ### Expected behavior Application no longer causes BSODs ### Full `tauri info` output ```text [โœ”] Environment - OS: Windows 10.0.19045 X64 โœ” WebView2: 130.0.2849.80 โœ” MSVC: Visual Studio Community 2022 โœ” rustc: 1.82.0 (f6e511eec 2024-10-15) โœ” cargo: 1.82.0 (8f40fc59f 2024-08-21) โœ” rustup: 1.27.1 (54dd3d00f 2024-04-24) โœ” Rust toolchain: stable-x86_64-pc-windows-msvc (default) - node: 21.7.1 - npm: 9.1.2 [-] Packages - tauri [RUST]: 1.8.1 - tauri-build [RUST]: 1.5.5 - wry [RUST]: 0.24.11 - tao [RUST]: 0.16.10 - @tauri-apps/api [NPM]: 1.6.0 (outdated, latest: 2.1.1) - @tauri-apps/cli [NPM]: 1.6.3 (outdated, latest: 2.1.0) [-] App - build-type: bundle - CSP: unset - distDir: ../dist - devPath: http://localhost:1420/ - framework: Vue.js - bundler: Vite ``` ### Stack trace _No response_ ### Additional context _No response_
type: bug,status: needs triage
low
Critical
2,676,713,163
angular
Select tag is flickering SSR
### Which @angular/* package(s) are the source of the bug? Don't known / other ### Is this a regression? Yes ### Description As you can see on the video, the select tag is flickering compared to input tag, I use SSR ### Please provide a link to a minimal reproduction of the bug https://stackblitz.com/~/github.com/illunix/flickering-issue-angular ### Please provide the exception or error you saw https://github.com/user-attachments/assets/71172d3c-bb6b-462b-8738-84b3aa88419e ### Please provide the environment you discovered this bug in (run `ng version`) ```true Angular CLI: 19.0.0 Node: 20.16.0 Package Manager: npm 10.9.0 OS: win32 x64 ``` ### Anything else? _No response_
workaround2: non-obvious,area: forms,area: server,P3
low
Critical
2,676,713,418
go
build: build failure on gotip-linux-loong64
``` #!watchflakes default <- builder == "gotip-linux-loong64" && repo == "go" && mode == "build" ``` Issue created automatically to collect these failures. Example ([log](https://ci.chromium.org/b/8730734133647122513)): [W2024-11-21T00:44:36.689641+08:00 1080691 0 json_subtract.go:121] Unknown fields while parsing property namespace "": {"is_google":false,"mode":0} 2024/11/21 00:44:36 run starting 2024/11/21 00:44:37 installed tools โ€” [watchflakes](https://go.dev/wiki/Watchflakes)
NeedsInvestigation
low
Critical
2,676,749,008
vscode
Quick search does not retain search term when moving to view
Notice the term isn't carried over to the view: ![Image](https://github.com/user-attachments/assets/f9185c18-257f-44b1-b29b-5a85dca47b33)
bug,search
low
Minor
2,676,749,907
pytorch
Spsolve for sparse, complex matrices
### ๐Ÿ› Describe the bug Sparse linear algebra matrix solver doesn't work with complex numbers on a GPU (solver not available at all on the CPU) Sample Code: ``` import torch from torch import sparse n = 10 A = torch.rand(n, n).type(torch.complex128).to_sparse_csr().cuda() b = torch.rand(n).type(torch.complex128).cuda() return sparse.spsolve(A, b) ``` Traceback: ``` File "test.py", line 36, in solve_matrix_equation return sparse.spsolve(A, b) RuntimeError: "create_matrix" not implemented for 'ComplexDouble' ``` ### Versions Collecting environment information... PyTorch version: 2.6.0a0+git99a0321 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Rocky Linux release 8.10 (Green Obsidian) (x86_64) GCC version: (GCC) 8.5.0 20210514 (Red Hat 8.5.0-22) Clang version: Could not collect CMake version: version 3.31.0 Libc version: glibc-2.28 Python version: 3.10.15 (main, Oct 3 2024, 07:27:34) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-4.18.0-553.8.1.el8_10.x86_64-x86_64-with-glibc2.28 Is CUDA available: True CUDA runtime version: 12.4.131 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: Tesla V100-SXM2-16GB Nvidia driver version: 555.42.06 cuDNN version: Probably one of the following: /usr/lib64/libcudnn.so.8.9.7 /usr/lib64/libcudnn_adv_infer.so.8.9.7 /usr/lib64/libcudnn_adv_train.so.8.9.7 /usr/lib64/libcudnn_cnn_infer.so.8.9.7 /usr/lib64/libcudnn_cnn_train.so.8.9.7 /usr/lib64/libcudnn_ops_infer.so.8.9.7 /usr/lib64/libcudnn_ops_train.so.8.9.7 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 8 On-line CPU(s) list: 0-7 Thread(s) per core: 2 Core(s) per socket: 4 Socket(s): 1 NUMA node(s): 1 Vendor ID: GenuineIntel CPU family: 6 Model: 79 Model name: Intel(R) Xeon(R) CPU E5-2686 v4 @ 2.30GHz Stepping: 1 CPU MHz: 2703.014 CPU max MHz: 3000.0000 CPU min MHz: 1200.0000 BogoMIPS: 4600.03 Hypervisor vendor: Xen Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 256K L3 cache: 46080K NUMA node0 CPU(s): 0-7 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single pti fsgsbase bmi1 hle avx2 smep bmi2 erms invpcid rtm rdseed adx xsaveopt Versions of relevant libraries: [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.26.4 [pip3] optree==0.13.1 [pip3] torch==2.6.0a0+git99a0321 [conda] blas 1.0 mkl [conda] magma-cuda124 2.6.1 1 pytorch [conda] mkl 2023.1.0 h213fc3f_46344 [conda] mkl-include 2025.0.1 pypi_0 pypi [conda] mkl-service 2.4.0 py310h5eee18b_1 [conda] mkl-static 2025.0.1 pypi_0 pypi [conda] mkl_fft 1.3.11 py310h5eee18b_0 [conda] mkl_random 1.2.8 py310h1128e8f_0 [conda] numpy 1.26.4 py310h5f9d8c6_0 [conda] numpy-base 1.26.4 py310hb5e798b_0 [conda] optree 0.13.1 pypi_0 pypi [conda] torch 2.6.0a0+git99a0321 pypi_0 pypi cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer @jcaip
module: sparse,triaged
low
Critical
2,676,763,197
vscode
Provide quick fix for missing copyright statement in CSS files
TS file: ![Image](https://github.com/user-attachments/assets/2be03b48-cc51-4f23-9ec8-5050c23791f3) CSS file: ![Image](https://github.com/user-attachments/assets/47b708ef-fa39-47bf-a4a9-1d283640628d)
feature-request,engineering
low
Minor
2,676,770,966
next.js
All imports must resolve in `next.config` when running `next lint`. This prevents monorepos from linting before building dependent packages.
### Link to the code that reproduces this issue https://codesandbox.io/p/devbox/damp-sea-q85m6v ### To Reproduce `npm run lint` ### Current vs. Expected behavior `Error [ERR_PACKAGE_PATH_NOT_EXPORTED]: No "exports" main defined in /project/sandbox/node_modules/some-package/package.json imported from /project/sandbox/next.config.mjs` It should be possible to lint a project _without_ a package being used as a plugin in the next config being built. In monorepo pipelines it's common to lint prior to building all dependent packages. Running `eslint` independently in a monorepo does not complain when a package referenced by a `file:` resolution in `package.json` points to non-existent files; this is a run-time not a lint-time concern. ### Provide environment information ```bash Operating System: Platform: linux Arch: x64 Version: #1 SMP PREEMPT_DYNAMIC Sun Aug 6 20:05:33 UTC 2023 Available memory (MB): 4102 Available CPU cores: 2 Binaries: Node: 20.9.0 npm: 9.8.1 Yarn: 1.22.19 pnpm: 8.10.2 Relevant Packages: next: 15.0.4-canary.20 // Latest available version is detected (15.0.4-canary.20). eslint-config-next: N/A react: 19.0.0-rc-380f5d67-20241113 react-dom: 19.0.0-rc-380f5d67-20241113 typescript: 5.3.3 Next.js Config: output: N/A ``` ### Which area(s) are affected? (Select all that apply) Developer Experience, Linting ### Which stage(s) are affected? (Select all that apply) next dev (local) ### Additional context _No response_
bug,Linting
low
Critical
2,676,780,132
flutter
Deprecate and remove Dart-side `DeferredComponents`
Flutter supports two sets of features around deferring code: (a) deferring chunks of Dart and (b) deferring the engine. The latter (b) is used, including in google3, to, quote _"not pay 5mb when a 1000 screen app adds their first 1 Flutter screen"_. The former feature, however, is not known to have any internal use, and we suspect very little (or no) external use. Feature overview: - https://github.com/flutter/flutter/wiki/Deferred-Components - https://docs.flutter.dev/perf/deferred-components From talking to @jonahwilliams and in my own experiences spelunking through the `flutter` tool, deferred components is a complex feature that is almost not tested at all, and has implementation elements throughout most of the product - in the tool, framework, and engine. Deprecating and removing it would remove a significant source of [haunted graveyard](https://www.usenix.org/sites/default/files/conference/protected-files/srecon17americas_slides_reese.pdf)-style complexity, especially if the feature is not used or tested. As a first step, I'll work with @andrewkolos to enable Google analytics tracking for this feature.
team,platform-android,c: API break,c: proposal,P2,c: tech-debt,fyi-android,fyi-tool
low
Minor
2,676,789,775
PowerToys
Screen Ruler Crashing
### Microsoft PowerToys version 0.86.0 ### Installation method PowerToys auto-update ### Running as admin None ### Area(s) with issue? Screen ruler ### Steps to reproduce STEP 1: - Go down to the Windows10 taskbar and click on the PowerToys icon. - Click on Screen Ruler from the quick access menu. OR - Use the default Win + Shift + m STEPs 2-4: - This itself doesn't always bring up the tool, but when it does press the Vertical Spacing button. - If it hasn't frozen yet, click anywhere on your screen. - Without fail it will freeze and become unresponsive requiring killing the process, ### โœ”๏ธ Expected Behavior For the tool to open every time I press the icon from the taskbar or use the keyboard shortcut and for the tool not to crash. ### โŒ Actual Behavior The tool inconsistently opens using any regular method of opening it and always crashes when it does manage to open within a very short amount of time (seemingly after one action and then clicking off). ### Other Software [PowerToysReport_2024-11-20-12-25-12.zip](https://github.com/user-attachments/files/17835262/PowerToysReport_2024-11-20-12-25-12.zip)
Issue-Bug,Needs-Triage
low
Critical
2,676,790,575
kubernetes
fix device manager permafailing test
### Which jobs are failing? https://testgrid.k8s.io/sig-node-presubmits#pr-node-kubelet-serial-containerd-flaky ### Which tests are failing? E2eNode Suite: [It] [sig-node] Device Plugin [NodeFeature:DevicePlugin] [Serial] DevicePlugin [Serial] [Disruptive] Keeps device plugin assignments across node reboots (no pod restart, no device plugin re-registration) [Flaky] ### Since when has it been failing? Oct 30 2024, likely before https://github.com/kubernetes/kubernetes/issues/128114 ### Testgrid link https://prow.k8s.io/view/gs/kubernetes-ci-logs/pr-logs/directory/pull-kubernetes-node-kubelet-containerd-flaky/1851562955540271104 ### Reason for failure (if possible) need to review the test purpose and update or remove it, with a rationale ### Anything else we need to know? followup of https://github.com/kubernetes/kubernetes/issues/128443 put issue on me to make sure we followup and we review the tests ### Relevant SIG(s) /sig
priority/important-soon,sig/node,kind/failing-test,triage/accepted
low
Critical
2,676,812,466
next.js
Inconsistent behaviour of revalidatePath between local and Vercel
### Link to the code that reproduces this issue https://github.com/damianfrizzi/next-revalidate-bug ### To Reproduce - Go to [https://next-revalidate-bug.vercel.app](https://t.co/BBukGxf9iY) and note the number at the top - Navigate to "Go to some page" and note the number at the top - Go back to homepage and click "Invalidate German" and refresh page - Number correctly changes on the homepage - BUT number also changes on the "/somePage" route. This shouldn't happen ### Current vs. Expected behavior `revalidatePath` behaves correctly when using `bun build && bun start` on my local machine. I.e., invalidating the homepage doesn't invalidate any other (sub) pages. When deployed to Vercel, invalidating the homepage also invalidates other (sub) pages. ### Provide environment information ```bash Operating System: Platform: darwin Arch: arm64 Version: Darwin Kernel Version 24.0.0: Tue Sep 24 23:39:07 PDT 2024; root:xnu-11215.1.12~1/RELEASE_ARM64_T6000 Available memory (MB): 32768 Available CPU cores: 10 Binaries: Node: 20.17.0 npm: 10.8.2 Yarn: 1.22.19 pnpm: N/A Relevant Packages: next: 15.0.3 // Latest available version is detected (15.0.3). eslint-config-next: 15.0.3 react: 19.0.0-rc-66855b96-20241106 react-dom: 19.0.0-rc-66855b96-20241106 typescript: 5.6.3 Next.js Config: output: N/A ``` ### Which area(s) are affected? (Select all that apply) Internationalization (i18n), Middleware, Runtime ### Which stage(s) are affected? (Select all that apply) Vercel (Deployed) ### Additional context I'm using `trailingSlash: true` in my next.config. I suspect that there is an error in connection with this configuration.
bug,Middleware,Internationalization (i18n),Runtime
low
Critical
2,676,884,649
go
crypto/ecdsa: Sign() panics if public key is not set
### Go version go version devel go1.24-7c7170e Wed Nov 20 18:27:31 2024 +0000 linux/amd64 ### Output of `go env` in your module/workspace: ```shell AR='ar' CC='clang-15' CGO_CFLAGS='-O2 -g' CGO_CPPFLAGS='' CGO_CXXFLAGS='-O2 -g' CGO_ENABLED='1' CGO_FFLAGS='-O2 -g' CGO_LDFLAGS='-O2 -g' CXX='clang++-15' GCCGO='gccgo' GO111MODULE='' GOAMD64='v1' GOARCH='amd64' GOAUTH='netrc' GOBIN='' GOCACHE='/home/jhg/.cache/go-build' GODEBUG='' GOENV='/home/jhg/.config/go/env' GOEXE='' GOEXPERIMENT='' GOFIPS140='off' GOFLAGS='' GOGCCFLAGS='-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -Wl,--no-gc-sections -fmessage-length=0 -ffile-prefix-map=/tmp/go-build1928511716=/tmp/go-build -gno-record-gcc-switches' GOHOSTARCH='amd64' GOHOSTOS='linux' GOINSECURE='' GOMOD='/home/jhg/oss-fuzz-379773077/p/go.mod' GOMODCACHE='/home/jhg/oss-fuzz-379773077/go-dev/packages/pkg/mod' GONOPROXY='' GONOSUMDB='' GOOS='linux' GOPATH='/home/jhg/oss-fuzz-379773077/go-dev/packages' GOPRIVATE='' GOPROXY='https://proxy.golang.org,direct' GOROOT='/home/jhg/oss-fuzz-379773077/go-dev' GOSUMDB='sum.golang.org' GOTELEMETRY='local' GOTELEMETRYDIR='/home/jhg/.config/go/telemetry' GOTMPDIR='' GOTOOLCHAIN='auto' GOTOOLDIR='/home/jhg/oss-fuzz-379773077/go-dev/pkg/tool/linux_amd64' GOVCS='' GOVERSION='devel go1.24-7c7170e Wed Nov 20 18:27:31 2024 +0000' GOWORK='' PKG_CONFIG='pkg-config' ``` ### What did you do? ```go package main import ( "crypto/ecdsa" "crypto/elliptic" "crypto/rand" "crypto/sha256" "fmt" "math/big" ) func main() { priv, ok := new(big.Int).SetString("123", 10) if ok == false { panic("Cannot decode bignum") } var priv_ecdsa ecdsa.PrivateKey priv_ecdsa.D = priv priv_ecdsa.PublicKey.Curve = elliptic.P256() msg := "hello, world" hash := sha256.Sum256([]byte(msg)) r, s, err := ecdsa.Sign(rand.Reader, &priv_ecdsa, hash[:]) fmt.Println(err) fmt.Println(r.String()) fmt.Println(s.String()) } ``` ### What did you see happen? ``` panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x10 pc=0x4c2117] goroutine 1 [running]: math/big.(*Int).Sign(...) /home/jhg/oss-fuzz-379773077/go-dev/src/math/big/int.go:48 crypto/ecdsa.pointFromAffine({0x56bc88?, 0x646070?}, 0x0, 0x0) /home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:402 +0x37 crypto/ecdsa.privateKeyToFIPS[...](0xc0002d02c0, 0xc0002dff18) /home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:391 +0x38 crypto/ecdsa.signFIPS[...](0xc0002d02c0, 0xc0002c8df8, {0x56b020?, 0x66c280}, {0xc0002dfed8, 0x20, 0x20}) /home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:234 +0x46 crypto/ecdsa.SignASN1({0x56b020, 0x66c280}, 0xc0002dff18, {0xc0002dfed8, 0x20, 0x20}) /home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:220 +0x229 crypto/ecdsa.Sign({0x56b020?, 0x66c280?}, 0xbf7cbadf074d6483?, {0xc0002c8ed8?, 0xc0002c8ef8?, 0xc?}) /home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa_legacy.go:60 +0x37 main.main() /home/jhg/oss-fuzz-379773077/p/x.go:23 +0xf4 ``` ### What did you expect to see? No panic and output like: ``` <nil> 40753909606936490861524166827361514810969838335424734688996005945565630866707 3003946056421339230760555414094068582110502790139132375823642300394845145720 ```
NeedsInvestigation
low
Critical
2,676,906,576
fastapi
Will FastAPI support QUERY http method? "app.query"
### Discussed in https://github.com/fastapi/fastapi/discussions/6049 <div type='discussions-op-text'> <sup>Originally posted by **FilipeMarch** December 16, 2022</sup> ### First Check - [X] I added a very descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the FastAPI documentation, with the integrated search. - [X] I already searched in Google "How to X in FastAPI" and didn't find any information. - [X] I already read and followed all the tutorial in the docs and didn't find an answer. - [X] I already checked if it is not related to FastAPI but to [Pydantic](https://github.com/samuelcolvin/pydantic). - [X] I already checked if it is not related to FastAPI but to [Swagger UI](https://github.com/swagger-api/swagger-ui). - [X] I already checked if it is not related to FastAPI but to [ReDoc](https://github.com/Redocly/redoc). ### Commit to Help - [X] I commit to help with one of those options ๐Ÿ‘† ### Example Code ```python from fastapi import FastAPI app = FastAPI() @app.query('/query/subjects') def query_subjects(schema: ArbitrarySchema): with Session(engine) as db: subjects = db.query(Subject).all() return schema(**subjects) # something like this ``` ### Description There is a new HTTP method called QUERY, I discovered it this week and it is super interesting! https://www.ietf.org/archive/id/draft-ietf-httpbis-safe-method-w-body-02.html I was trying to make a GET route that would accept a request body, but started receiving an error on Swagger UI: [TypeError: Request has method 'GET' and cannot have a body](https://github.com/swagger-api/swagger-ui/issues/5891#issuecomment-1340591857) The idea is that sometimes we need to make a big or complex query, and this is the scenario in which we can see the advantages of GraphQL, and although I have seen that [we can integrate FastAPI with GraphQL](https://fastapi.tiangolo.com/advanced/graphql/), I was wondering if FastAPI will ever be able to simply accept a QUERY method like, for example, `app.query("/query/subjects")`. For example, suppose I have this object: ```yml { "id": 1, "name": "Math" "tags": [ { "id": 1, "name": "Algebra", "number_of_clicks": 1, "number_of_questions": 7, "number_of_answers": 3, "number_of_comments": 2, "number_of_votes": 1, }] "topics": [ { "id": 1, "name": "Linear Equations", "likes": 1, "dislikes": 0, "number_of_clicks": 1, "number_of_tutorials": 1, "number_of_questions": 7, "posts": [ { "id": 1, "title": "How to solve linear equations?", "likes": 1, "dislikes": 0, "number_of_clicks": 1, "number_of_answers": 3, "number_of_comments": 2, "number_of_votes": 1, "answers": [ { "id": 1, "content": "You can solve linear equations by using the substitution method.", "likes": 1, "dislikes": 0, "number_of_clicks": 1, "number_of_comments": 2, "number_of_votes": 1, "comments": [ { "id": 1, "content": "That's a great answer!", "likes": 1, "dislikes": 0, "number_of_clicks": 1, "number_of_votes": 1, }" ] } ] } ] } ] } ``` I want the client to be able to QUERY this in any specific way, maybe he wants GET objects with this format ```yml { "id": 1, "name": "Math" "tags": [ { "id": 1, "name": "Algebra", }] "topics": [ { "id": 1, "name": "Linear Equations", } ] } ``` But it could be totally different, and I can't predict the way the client is gonna query it and make a new route for every combination and possible schema for a complex object. The QUERY route would accept a request body, and then we would filter the response with the schema the client sends on request body. - Would it be too complicated to implement this on FastAPI? - Is there any hope that this will someday be implemented in FastAPI? XD Thanks! ### Wanted Solution ... ### Wanted Code ```python ... ``` ### Alternatives ... ### Operating System macOS ### Operating System Details _No response_ ### FastAPI Version A ### Python Version 3.11 ### Additional Context _No response_</div>
feature,good first issue
low
Critical
2,676,962,886
vscode
Feature Request: Expose findWidget Actions State via API
The VS Code Find Widget allows toggling control options like `toggleFindCaseSensitive`, `toggleFindWholeWord`, `toggleFindRegex`. However, there is no way to programmatically query the current state of these options via the extension API (that I'm aware of). This limitation prevents extensions from seamlessly integrating with or enhancing the Find Widgetโ€™s functionality. ### Proposed Solution Expose the current state of the Find Widget toggles (e.g., `findCaseSensitive`, `findWholeWord`, `findRegex`) through the extension API. These could be added to the `activeTextEditor` options either nested or directly: `vscode.window.activeTextEditor.options.findWidget.findCaseSensitive` An alternative would be to allow toggling to an explicit state: `vscode.commands.executeCommand("toggleFindCaseSensitive", true)` ### Workarounds / Context I've thus far not been able to find a way to work around this- I am attempting to write an extension that modifies the cmd+d selection functionality to more closely match Atom/Pulsar in such a way that it determines if you are trying to select a full word (cursor on the word, nothing actively selected) vs partial matches (highlighted/selected text first, THEN cmd+d) but because there is no way to check these settings I am unable to toggle them to the desired state based on context.
feature-request,api
low
Minor
2,676,968,806
deno
Upgrade to TypeScript 5.7
We're going to push this back a release to Deno 2.2 due to https://devblogs.microsoft.com/typescript/announcing-typescript-5-7-beta/#typedarrays-are-now-generic-over-arraybufferlike in order to find a better migration strategy.
feat,tsc
low
Minor
2,677,039,656
ui
[bug]: Tooltip on Themes page doesn't inherit theme
### Describe the bug On the Themes page: https://ui.shadcn.com/themes When hovering over the new message button in the ChatDemo, the Tooltip portals in on the body, which is outside the .theme-red selector so the tooltip doesn't inherit the theme. I think one potential solution could be to change the tooltip.tsx component to portal the tooltip content in relative to the tooltip root vs. on document.body. ### Affected component/components Documentation, Tooltip ### How to reproduce 1. Go to https://ui.shadcn.com/themes 2. Hover on new message button in ChatDemo 3. Notice tooltip is black, when it should be red if the red theme is selected. ### Codesandbox/StackBlitz link https://ui.shadcn.com/themes ### Logs _No response_ ### System Info ```bash Mac, Chrome ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,677,041,726
PowerToys
Cannot launch UI elements: faulting module Microsoft.UI.Xaml.dll
### Microsoft PowerToys version 0.86.0 ### Installation method GitHub ### Running as admin No ### Area(s) with issue? General, Settings, Advanced Paste, PowerToys Run ### Steps to reproduce Installed 0.86.0 (machine-wide) from GitHub on W11 23H2 22631 (multiple machines), both manually and through SCCM. I get the following error when trying to launch any UI part of PowerToys: ``` Faulting application name: PowerToys.PowerLauncher.exe, version: 0.86.0.0, time stamp: 0x66e80000 Faulting module name: KERNELBASE.dll, version: 10.0.22621.4391, time stamp: 0x7433a115 Exception code: 0xe0434352 Fault offset: 0x000000000005fa4c Faulting process ID: 0x0xbb0 Faulting application start time: 0x0x1db3b83c903a38c Faulting application path: C:\Program Files\PowerToys\PowerToys.PowerLauncher.exe Faulting module path: C:\WINDOWS\System32\KERNELBASE.dll ``` ``` Faulting application name: PowerToys.Settings.exe, version: 0.86.0.0, time stamp: 0x66e80000 Faulting module name: Microsoft.UI.Xaml.dll, version: 3.1.6.0, time stamp: 0x0caad657 Exception code: 0xc000027b Fault offset: 0x0000000000009125 Faulting process ID: 0x0xf24 Faulting application start time: 0x0x1db3b83caefa763 Faulting application path: C:\Program Files\PowerToys\WinUI3Apps\PowerToys.Settings.exe Faulting module path: C:\Program Files\PowerToys\WinUI3Apps\Microsoft.UI.Xaml.dll ``` ``` Faulting application name: PowerToys.Peek.UI.exe, version: 0.86.0.0, time stamp: 0x66e80000 Faulting module name: Microsoft.UI.Xaml.dll, version: 3.1.6.0, time stamp: 0x0caad657 Exception code: 0xc000027b Fault offset: 0x0000000000009125 Faulting process ID: 0x0x1b20 Faulting application start time: 0x0x1db3b84ec20d9a2 Faulting application path: C:\Program Files\PowerToys\WinUI3Apps\PowerToys.Peek.UI.exe Faulting module path: C:\Program Files\PowerToys\WinUI3Apps\Microsoft.UI.Xaml.dll ``` --- Trying to generate a debug report results in (likely because diagnostics is disabled through GPO): ``` Application: PowerToys.PowerLauncher.exe CoreCLR Version: 8.0.1024.46610 .NET Version: 8.0.10 Description: The process was terminated due to an unhandled exception. Exception Info: System.InvalidCastException: Unable to cast object of type 'System.Byte[]' to type 'System.Int32'. at Microsoft.PowerToys.Telemetry.DataDiagnosticsSettings.GetEnabledValue() at Microsoft.PowerToys.Telemetry.ETWTrace..ctor() at PowerLauncher.App..ctor() at PowerLauncher.App.Main() ``` ### โœ”๏ธ Expected Behavior Any UI element of PowerToys opens as expected. ### โŒ Actual Behavior Nothing opens (the icon is visible in the system tray), and occasionally, the system hangs (especially when trying to do something with advanced paste). Interestingly, Always On Top works perfectly. Running the application as an administrator also works perfectly (no GPO enforced). ### Other Software _No response_
Issue-Bug,Resolution-Fix Committed
medium
Critical
2,677,049,900
react
[React 19] Regression when using `createPortal` with DOM element created by `dangerouslySetInnerHTML`
In React 18, it was possible to use `createPortal` with a DOM element created by `dangerouslySetInnerHTML`. Example (adapted from [this Stack Overflow answer](https://stackoverflow.com/questions/62300569/add-react-component-in-html-set-by-dangerouslysetinnerhtml/77099814#77099814)): ```tsx import { useCallback, useState } from "react"; import { createPortal } from "react-dom"; export default function App() { const htmlFromElsewhere = `foo <span class="portal-container"></span> bar`; return <InnerHtmlWithPortals html={htmlFromElsewhere} /> } function InnerHtmlWithPortals({ html }: { html: string }) { const [portalContainer, setPortalContainer] = useState<Element | null>(null) const refCallback = useCallback((el: HTMLDivElement) => { setPortalContainer(el?.querySelector(".portal-container")) }) return <> <div ref={refCallback} dangerouslySetInnerHTML={{ __html: html }} /> {portalContainer && createPortal(<Cake />, portalContainer)} </> } function Cake() { return <strong>cake</strong> } ``` React 18 CodeSandbox: https://codesandbox.io/p/sandbox/optimistic-kowalevski-73sk5w In React 19, this no longer works. React appears to be re-rendering the inner HTML after calling `refCallback`. Thus, `createPortal` succeeds, but the `portalContainer` element that it uses is no longer part of the DOM. React 19 CodeSandbox: https://codesandbox.io/p/sandbox/vibrant-cloud-gd8yzr It is possible to work around the issue by setting `innerHTML` directly instead of using `dangerouslySetInnerHTML`: ```diff const [portalContainer, setPortalContainer] = useState<Element | null>(null) const refCallback = useCallback((el: HTMLDivElement) => { + if (el) el.innerHTML = html setPortalContainer(el?.querySelector(".portal-container")) - }) + }, [html]) return <> - <div ref={refCallback} dangerouslySetInnerHTML={{ __html: html }} /> + <div ref={refCallback} /> {portalContainer && createPortal(<Cake />, portalContainer)} </> ``` But I'm not sure whether that is a reliable solution.
React 19
medium
Minor
2,677,073,607
kubernetes
[FG:InPlacePodVerticalScaling] pull-kubernetes-e2e-capz-windows-master test fail with InPlacePodVerticalScaling Beta
### Which jobs are failing? pull-kubernetes-e2e-capz-windows-master https://prow.k8s.io/view/gs/kubernetes-ci-logs/pr-logs/pull/128880/pull-kubernetes-e2e-capz-windows-master/1859240005449289728 ### Which tests are failing? Kubernetes e2e suite: [It] [sig-storage] - Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] -Secrets should be consumable from pods in volume [NodeConformance] [Conformance] -ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 23s -Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 33s -Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 25s - Projected secret should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] expand_more 25s - Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] expand_more 27s - Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] expand_more 31s - Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] expand_more 27s - Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 27s - Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] expand_more 25s - Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] expand_more 27s - ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] expand_more 1m9s - ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] expand_more 25s - ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] expand_more 27s - Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] expand_more Kubernetes e2e suite: [It] [sig-windows] [Feature:Windows] - Windows volume mounts check volume mount permissions container should have readOnly permissions on emptyDir expand_more 35s Kubernetes e2e suite: [It] [sig-node] - Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] expand_more 10m3s ### Since when has it been failing? Last successful 12 November 20:20 1856416666905219072 https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-capz-master-windows/1856416666905219072 First failed 12 November 23:20 1856461965182898176 https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-capz-master-windows/1856461965182898176 ### Testgrid link https://testgrid.k8s.io/sig-release-master-informing#capz-windows-master ### Reason for failure (if possible) _No response_ ### Anything else we need to know? _No response_ ### Relevant SIG(s) /sig-node /sig-storage /sig-windows
priority/important-soon,sig/node,sig/windows,kind/failing-test,triage/accepted
medium
Critical
2,677,080,060
go
proposal: hash: standardize the hash function
**Background**: Issue https://github.com/golang/go/issues/69420 proposes a `types.Hash` function that provides a hash operator for `types.Type` values that is consistent with the equivalence relation for types defined by `types.Identical`. The only real question in that proposal is: what is to be Go's future convention for the signature of a custom hash function? The naive answer is `func(T) uint64`, where T stands for `types.Type`. However, hash tables with only a single hash function are vulnerable to a "hash flooding" denial-of-service attack, in which an adversary provides many inputs that all hash to the same value, causing the table to degenerate into a linked list. The defense is for each process (or particular hash table) to select a different hash function at random so that the values cannot be predicted. The `hash/maphash` package embodies this approach: a random Seed value determines the hash functions (`Hash.Strings` and `Hash.Bytes`), which depend on the seed's opaque random value. (maphash does not provide a Seed-dependent way to hash integers.) So, perhaps custom hash functions should accept a `maphash.Seed` parameter. And perhaps Seed should also provide methods for hashing integers. **Proposal:** In the hash package, define function types HashFunc and EqFunc that document the conventions for Go hash tables. Here is the minimal answer, without Seeds: ```go package hash // HashFunc defines a hash function for values of type T. // A conformant HashFunc and EqFunc pair (hash, eq) must observe the invariant // that eq(x, y) => hash(x) == hash(y). type HashFunc[T any] func(T) uint64 // EqFunc defines an equivalence relation for values of type T. type EqFunc[T any] func(x, y T) bool ``` @ianlancetaylor @griesemer
Proposal
medium
Major
2,677,083,268
pytorch
Dynamo should reorder warnings (and other logging) by default if they don't involve tensor values
The thing I ran into is https://gist.github.com/zou3519/4c45db951c17845a598f4e9d9f433ba9 (instancenorm warns if an input looks wrong). One can also run into this via deprecation warnings. cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @chauhang @amjames
triaged,oncall: pt2,module: dynamo,empathy-day
low
Minor
2,677,127,597
rust
Tracking Issue for OsString::truncate
<!-- Thank you for creating a tracking issue! Tracking issues are for tracking a feature from implementation to stabilization. Make sure to include the relevant RFC for the feature if it has one. If the new feature is small, it may be fine to skip the RFC process. In that case, you can use `issue = "none"` in your initial implementation PR. The reviewer will ask you to open a tracking issue if they agree your feature can be added without an RFC. --> Feature gate: `#![feature(os_string_truncate)]` This is a tracking issue for the <!-- Include a short description of the feature. --> ### Public API <!-- For most library features, it'd be useful to include a summarized version of the public API. (E.g. just the public function signatures without their doc comments or implementation.) --> ```rust // std::ffi::os_str impl OsString { pub fn truncate(&mut self, new_len: usize); } ``` ### Steps / History <!-- For larger features, more steps might be involved. If the feature is changed later, please add those PRs here as well. --> - [x] ACP Approved: https://github.com/rust-lang/libs-team/issues/486 - [ ] Implementation: - [ ] Final comment period (FCP)[^1] - [ ] Stabilization PR <!-- Once the feature has gone through a few release cycles and there are no unresolved questions left, the feature might be ready for stabilization. If this feature didn't go through the RFC process, a final comment period (FCP) is always needed before stabilization. This works as follows: A library API team member can kick off the stabilization process, at which point the rfcbot will ask all the team members to verify they agree with stabilization. Once enough members agree and there are no concerns, the final comment period begins: this issue will be marked as such and will be listed in the next This Week in Rust newsletter. If no blocking concerns are raised in that period of 10 days, a stabilization PR can be opened by anyone. --> ### Unresolved Questions <!-- Include any open questions that need to be answered before the feature can be stabilised. If multiple (unrelated) big questions come up, it can be a good idea to open a separate issue for each, to make it easier to keep track of the discussions. It's useful to link any relevant discussions and conclusions (whether on GitHub, Zulip, or the internals forum) here. --> - None yet. [^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html
T-libs-api,C-tracking-issue
low
Minor
2,677,131,714
flutter
Revisit inline vs out-of-line implementations of methods in Impeller geometry classes
Currently many of the Impeller geometry classes are templates and so they must implement all of their methods inline, but some of them (for instance `Matrix`) are regular classes and so the methods could be implemented in their associated `.cc` file. We've been shifting the code to more cases of moving implementations to the `.cc` file and so we should revisit the various geometry classes to determine what's best for them.
engine,P3,e: impeller,team-engine,triaged-engine
low
Minor
2,677,146,162
next.js
URL's search params get encoded on page load.
### Link to the code that reproduces this issue https://github.com/wkd-kapsule/next-router-demo ### To Reproduce 1. Go to the [production site.](https://next-router-demo-eight.vercel.app/) 2. Use a special character in the URL's search params, such as `/`. Whether by directly typing it or clicking the button on the page. 3. Watch the URL being encoded before your eyes as the page loads. ### Current vs. Expected behavior ### **_Current behavior:_** Directly typing a url or using `useRouter` will encode the URL's search params after page loading. ### **_Expected behavior:_** I'd like to be able to use characters such as `/` or `$` in a URL's search params. ### **_Context:_** I use URL seach params to store some state in my app. The thing is that it's a path, thus containing `/`. In dev mode, everything is fine but somehow in production, `/` gets encoded into `%2F` once the page has loaded. It does it whether I directly type the url and press Enter or click on button with `useRouter().push`. I've reproduced this in a minimal repo but for some reason it works fine when using `useRouter().push`. The URL gets encoded only when you manually type it and press Enter. There is basically no difference in the way I use `useRouter().push()` in this repo and in my app, I can't grasp why both behaves differently. **NOTE: I know that I can just decode it to get the original character but my issue is more visual than technical. Having encoded characters could discourage the users of my app to use the URL to navigate. Whereas this is exactly why I put that path in the URL, so the user can directly go where they need, without having to use the interface.** Here's the [demo](https://next-router-demo-eight.vercel.app/) https://github.com/user-attachments/assets/4d63d9d5-f00f-4c84-b6ef-19600627b2b3 ### Provide environment information ```bash Operating System: Platform: win32 Arch: x64 Version: Windows 11 Home Available memory (MB): 16272 Available CPU cores: 8 Binaries: Node: 20.13.1 npm: 10.8.1 Yarn: N/A pnpm: 8.15.5 Relevant Packages: next: 15.0.4-canary.19 // There is a newer canary version (15.0.4-canary.20) available, please upgrade! eslint-config-next: 15.0.3 react: 19.0.0-rc-69d4b800-20241021 react-dom: 19.0.0-rc-69d4b800-20241021 typescript: 5.6.2 Next.js Config: output: N/A ``` ### Which area(s) are affected? (Select all that apply) Navigation ### Which stage(s) are affected? (Select all that apply) Vercel (Deployed) ### Additional context I couldn't pin which Next.js versions first caused the issue because it actually occured a while ago. I didn't report it at the time as being constantly on the latest canary version, I just imagined it would eventually get solved. Now that's been several months already so I thought I better report it. Nonetheless, I can tell you that I'm using Google Chrome and that everything works as expected in Microsoft Edge.
bug,Navigation
low
Minor
2,677,200,814
ollama
Request: Nexa AI Omnivision
Here is the link to Nexa AI's model. (I haven't checked over it to make sure it's super reputable though) https://huggingface.co/NexaAIDev/omnivision-968M
model request
low
Major
2,677,208,897
ant-design
`RangePicker` breaks when using with date-fns via `dateFnsGenerateConfig`
### Reproduction link [![Edit on CodeSandbox](https://codesandbox.io/static/img/play-codesandbox.svg)](https://codesandbox.io/p/sandbox/datepicker-rangepicker-antd-v5-hr2rr9?file=%2Findex.tsx) ### Steps to reproduce - Setup Antd DatePicker to use date-fns according the [official guide](https://ant.design/docs/react/use-custom-date-library#use-date-fns); - Renders `RangePicker` using `format`, `picker` containing `"time"` value and `use12Hours` props; - Click on the RangePicker to input values. ### What is expected? The RangePicker should show the dropdown with the options so we can select a time range. ### What is actually happening? The entire app breaks with an error "Format string contains an unescaped latin alphabet character `A`" | Environment | Info | | --- | --- | | antd | 5.22.1 | | React | ^18.0.0 | | System | macOS Sequoia Version 15.0 (24A335) | | Browser | Google Chrome Version 131.0.6778.86 | --- I walked through the previous versions to find when this starts to break and I found the `v5.18.3` working but if we upgrade to `v5.19.0` or earlier, the error starts to happen. <!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
help wanted,Inactive
low
Critical
2,677,214,240
ui
[bug]: GitHub link in site header is only clickable on icon
### Describe the bug The GitHub link in the site header is only clickable on the icon, if the button is clicked outside of the icon, the link doesn't work. This is solved by using asChild as documented in https://ui.shadcn.com/docs/components/button ### Affected component/components Button, Link ### How to reproduce 1. Go to https://ui.shadcn.com/ 2. Click on the GitHub link in the site header, clicking outside of the icon 3. Doesn't work ### Codesandbox/StackBlitz link _No response_ ### Logs _No response_ ### System Info ```bash Any system ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,677,214,306
flutter
Removing the `Key` class
### Document Link [flutter.dev/go/unwrapping-widget-keys](https://docs.google.com/document/d/1Fc9ZYTTxlJUAfGeQzf83wXuV-wSiRZFHhMgwoUNukhI) ### What problem are you solving? This is an effort to make widgets less complicated and more performant.
c: API break,a: first hour,c: tech-debt,design doc,:scroll:
low
Minor
2,677,221,233
neovim
Windows WSL: osc52 data with CRLF lines gets Chinese chars
### Problem osc52 works for one line only. More than one line means bug (random chinese character etc.) ### Steps to reproduce nvim --clean -u minimal.lua write deed de select previous text press "y" your clipboard now is chinese ๆ•คๆ‘ฅๆ€e note on my full config (not just the minimal.lua) copying a selection always copy only the first line instead of becoming chinese clipboard. Also copying one line only always works flawlessly minimal.lua ```lua vim.g.clipboard = { name = 'OSC 52', copy = { ['+'] = require('vim.ui.clipboard.osc52').copy('+'), ['*'] = require('vim.ui.clipboard.osc52').copy('*'), }, paste = { ['+'] = require('vim.ui.clipboard.osc52').paste('+'), ['*'] = require('vim.ui.clipboard.osc52').paste('*'), }, } vim.cmd("set clipboard=unnamedplus") ``` ### Expected behavior Not chinese characters ### Nvim version (nvim -v) 0.10.2 ### Vim (not Nvim) behaves the same? vim no osc52 native AFAIK ### Operating system/version windows wsl ubuntu ### Terminal name/version windows terminal ### $TERM environment variable on ssh xterm-256color ### Installation appimage 0.10.2 github
bug,platform:windows,clipboard,needs:repro,platform:wsl
low
Critical
2,677,333,920
node
Support `concurrency` when `--experimental-test-isolation` is set to `'none'`
### What is the problem this feature will solve? It would be nice to be able to run multiple "workers" when using the `none` test isolation mode. Today, according to the docs, when `--experimental-test-isolation` is set to `'none'`, it implies 1 concurrency: https://github.com/nodejs/node/blob/f270462c09ddfd770291a7c8a2cd204b2c63d730/doc/api/cli.md#L2252-L2264 However, `mocha` another test runner that does not isolate tests, does accept a concurrency flag: https://mochajs.org/#parallel-tests ### What is the feature you are proposing to solve the problem? We're looking to move off of `jest` because its test module isolation is extremely slow. We love the idea of using the node-native test runner with isolation disabled, instead of adopting another third-party framework like `mocha`. However, we'd need to write some custom code (e.g., using [`parallel`](https://www.gnu.org/software/parallel/)) to spin up `n` concurrent, isolation-disabled tests to effectively utilize all the cores available on our CI machine. The docs and other recent comments all indicate that when `--experimental-test-isolation` is set to `'none'`, concurrency must be `1`. However, I couldn't find the reasoning in [the original PR](https://github.com/nodejs/node/pull/53927) or [issue](https://github.com/nodejs/node/issues/51548). There's probably a good reason for this but, as someone not intimately familiar with the implementation, it's not obvious to me why we wouldn't be able to run non-isolated tests concurrently, like in `mocha`. ### What alternatives have you considered? I could probably use a tool like [`parallel`](https://www.gnu.org/software/parallel/) to spin up multiple calls to `node --test`. However, this would require me to also write code to split up all the test files between the parallel runs, etc.
feature request,test_runner
low
Major
2,677,337,079
pytorch
cudagraph_trees error message suggestion doesn't always work
This came from running a user empathy model. Repro is: ``` import torch class Mod(torch.nn.Module): def __init__(self): super().__init__() def forward(self, x): out = x * 2 if not hasattr(self, "t"): self.t = torch.arange(x.shape[1], device='cuda')[None, None] elif self.t.shape[-1] < x.shape[1]: self.t = torch.arange(x.shape[1], device='cuda')[None, None] else: self.t = self.t[..., :x.shape[1]] return out m = Mod() m = torch.compile(m, mode='reduce-overhead') x = torch.randn(8, 8, device='cuda') with torch.no_grad(): torch.compiler.cudagraph_mark_step_begin() m(x) torch.compiler.cudagraph_mark_step_begin() m(x) torch.compiler.cudagraph_mark_step_begin() m(x) ``` This gives the error: ``` File "/home/hirsheybar/.conda/envs/pytorch_nightly/lib/python3.11/site-packages/torch/_inductor/cudagraph_trees.py", line 628, in get_non_cudagraph_inps and t.untyped_storage().data_ptr() not in existing_path_data_ptrs ^^^^^^^^^^^^^^^^^^^ RuntimeError: Error: accessing tensor output of CUDAGraphs that has been overwritten by a subsequent run. Stack trace: File "/data/users/hirsheybar/nightly/evo/tmp2.py", line 10, in forward self.t = torch.arange(x.shape[1], device='cuda')[None, None]. To prevent overwriting, clone the tensor outside of torch.compile() or call torch.compiler.cudagraph_mark_step_begin() before each model invocation. ``` cc @mcarilli @ezyang @eellison @penguinwu @chauhang @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @aakhundov
triaged,module: cuda graphs,oncall: pt2,module: inductor,empathy-day
low
Critical
2,677,359,945
excalidraw
[Feature Request] - Add a document to the canvas
I think a very good addition to Excalidraw would be to add a "Document" feature. Either the way Whimsical does it, where you can create documents separated from canvas, or the way Eraser does it, with a document attached to the canvas itself.
enhancement
low
Minor
2,677,368,559
angular
Can't build adev (bazel error)
### Which @angular/* package(s) are the source of the bug? Don't known / other ### Is this a regression? Yes ### Description I get this weird error from bazel when I try build adev ### Please provide a link to a minimal reproduction of the bug I followed every step here https://github.com/angular/angular/blob/main/adev/README.md, also tried cleaning after error but did not help ### Please provide the exception or error you saw ```true ERROR: C:/users/cros4/desktop/angular/packages/zone.js/BUILD.bazel:19:8: Executing genrule //packages/zone.js:LICENSE_wrapped failed: (Exit -1): bash.exe failed: error executing command c:\tools\msys64\usr\bin\bash.exe -c ... (remaining 1 argument skipped) Action failed to execute: java.io.IOException: ERROR: src/main/native/windows/process.cc(202): CreateProcessW("c:\tools\msys64\usr\bin\bash.exe" -c "source external/bazel_tools/tools/genrule/genrule-setup.sh; (echo '/** @license' && cat LICENSE && echo '*/') > bazel-out/x64_windows-fastbuild/bin/packages/zone.js/LICENSE.wrapped"): Can't find specified file. (error: 2) Target //adev:serve failed to build ``` ### Please provide the environment you discovered this bug in (run `ng version`) ```true OS: win32 x64 ``` ### Anything else? _No response_
area: docs-infra
low
Critical
2,677,369,816
pytorch
`CrossEntropyLoss` much slower with class probabilities vs. class indices target
### ๐Ÿ› Describe the bug Hi, I noticed this when training ViT-S/16 on the ImageNet-1k dataset, with and without [MixUp](https://arxiv.org/abs/1710.09412). In short: torchvision's [`v2.MixUp`](https://pytorch.org/vision/main/generated/torchvision.transforms.v2.MixUp.html), just like other common implementations, converts the class indices to one-hot class probabilities before linearly interpolate them, so the same call to `CrossEntropyLoss()` goes to different kernels with vs. without MixUp. In a less optimized setting ([specific commit](https://github.com/EIFY/mup-vit/tree/51bcdc77ea6e26ceef049e96f59702a9d9ef15d1)), I found that with vs. without MixUp makes the difference between **0.198s vs. 0.120s** per step (excluding data loading) with batch size 1024 on a 8x A100-SXM4-40GB [Lambda](https://lambdalabs.com/) instance. In further iterations, I included `CrossEntropyLoss` in the model so they compile together. I then tested a workaround in which the custom MixUp returns both targets and the weight scalar lam(bda): ```python lam = self._dist.sample() images = images.roll(1, 0).mul_(1.0 - lam).add_(images, alpha=lam) return images, lam, labels, labels.roll(1, 0) ``` and the model calculates the cross entropy loss by using the class indices twice: ```python logprob = F.log_softmax(out, dim=1) return lam * F.nll_loss(logprob, target1) + (1.0 - lam) * F.nll_loss(logprob, target2) ``` See https://github.com/EIFY/mup-vit/compare/normal-prefetch...loss-optim. I found that this is faster (**0.184s** per step) than just calling `CrossEntropyLoss()` with class probabilities even with `torch.compile(...,mode="max-autotune-no-cudagraphs")` (**0.191s** per step), excluding data loading. Is this expected? Intuitively I can see that the sparsity helps forward & backward pass in loss calculation, but backward pass beyond that should be the same and it's hard to understand the observed second/step difference with a 12-layer transformer. To replicate the timing result, you can run the code with fake data: ```shell MUPVIT_MAIN=~/Downloads/mup-vit/main.py PYTHON=torchrun N_WORKERS=80 N_THREADS=124 EPOCH=90 NUMEXPR_MAX_THREADS=$N_THREADS $PYTHON $MUPVIT_MAIN --fake-data --workers $N_WORKERS --multiprocessing-distributed --epochs $EPOCH --batch-size 1024 --mlp-head --report-to wandb --name cross-entropy-loss-class-prob NUMEXPR_MAX_THREADS=$N_THREADS $PYTHON $MUPVIT_MAIN --fake-data --workers $N_WORKERS --multiprocessing-distributed --epochs $EPOCH --batch-size 1024 --mlp-head --mixup-alpha 0.0 --report-to wandb --name cross-entropy-loss-class-index ``` with either the [normal-prefetch](https://github.com/EIFY/mup-vit/tree/normal-prefetch) or the [loss-optim](https://github.com/EIFY/mup-vit/tree/loss-optim) branch. ### Versions The output below shows PyTorch version: 2.3.1 and torchvision==0.18.1, but I am actually running ``` $ pip freeze (...) torch==2.5.1 torchaudio==2.5.1 torchvision==0.20.1 (...) ``` I don't know why `python3 -mpip list --format=freeze` finds the old packages. ``` Collecting environment information... PyTorch version: 2.3.1 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-6.8.0-47-generic-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: 12.4.131 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB GPU 1: NVIDIA A100-SXM4-40GB GPU 2: NVIDIA A100-SXM4-40GB GPU 3: NVIDIA A100-SXM4-40GB GPU 4: NVIDIA A100-SXM4-40GB GPU 5: NVIDIA A100-SXM4-40GB GPU 6: NVIDIA A100-SXM4-40GB GPU 7: NVIDIA A100-SXM4-40GB Nvidia driver version: 550.90.12 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 48 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 124 On-line CPU(s) list: 0-123 Vendor ID: AuthenticAMD Model name: AMD EPYC 7542 32-Core Processor CPU family: 23 Model: 49 Thread(s) per core: 1 Core(s) per socket: 1 Socket(s): 124 Stepping: 0 BogoMIPS: 5800.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities Virtualization: AMD-V Hypervisor vendor: KVM Virtualization type: full L1d cache: 7.8 MiB (124 instances) L1i cache: 7.8 MiB (124 instances) L2 cache: 62 MiB (124 instances) L3 cache: 1.9 GiB (124 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-31,64-95 NUMA node1 CPU(s): 32-63,96-123 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Mitigation; untrained return thunk; SMT disabled Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] flake8==4.0.1 [pip3] numpy==1.21.5 [pip3] optree==0.12.1 [pip3] torch==2.3.1 [pip3] torchvision==0.18.1 [pip3] triton==2.3.1 [conda] Could not collect ``` cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki
module: nn,module: loss,triaged
low
Critical
2,677,437,085
go
proposal: cmd/{go,cgo,compile}: change interpretation of relative //line directives
**Background**: generated Go source files may contain //line directives that relate each line of Go source to the line of some other file from which it was generated. If the file named by the line directive is a relative path, cmd/compile currently interprets it like any other relative path, that is, relative to its own os.Getwd. This demands that the tool that generated the //line directive has the same working directory as the compiler (e.g. GOROOT/src), or that it creates the line directive in anticipation of the consuming process having (say) GOROOT/src as its working directory. By contrast, go/scanner takes a different approach: it interprets relative line directives relative to the scanned file's directory. So, if it encounters the line directive `//line a.yacc:1` in a file named `a/a.go`, the relative line directive is resolved as `a/a.yacc`. (In March 2018 go/scanner was changed to resemble the compiler in [CL 100235](https://go.dev/cl/100235), but this was reverted in August 2018 by [CL 127658](https://go.dev/cl/127658).) The difference in behavior is easily observed by introducing a dummy line directive with a compiler error into an arbitrary Go source file. In this case, we'll append a relative reference from gopls/main.go to its README.md file: ```diff xtools$ git diff diff --git a/gopls/main.go b/gopls/main.go index 083c4efd8..b63784ba8 100644 --- a/gopls/main.go +++ b/gopls/main.go @@ -34,3 +34,6 @@ func main() { ctx := context.Background() tool.Main(ctx, cmd.New(), os.Args[1:]) } + +//line README.md:1 +invalid ``` When we build the package, the error message names a non-existent file: ``` xtools$ go build -o /dev/null ./gopls # golang.org/x/tools/gopls README.md:1: syntax error: non-declaration statement outside function body ``` When we use tools based on go/scanner, the error message names the correct file: ``` xtools$ gopackages -mode=allsyntax -deps ./gopls 2>&1 | grep README /Users/adonovan/w/xtools/gopls/README.md:1: expected declaration, found invalid ``` This is true even when we run the same commands from a different directory. Again, `go build` names the non-existent file, and go/scanner names the correct file: ``` xtools$ cd internal/ internal$ go build -o /dev/null ../gopls # golang.org/x/tools/gopls README.md:1: syntax error: non-declaration statement outside function body ``` ``` internal$ gopackages -mode=allsyntax -deps ../gopls 2>&1 | grep README /Users/adonovan/w/xtools/gopls/README.md:1: expected declaration, found invalid ``` **Proposal:** The compiler's scanner should interpret relative line directives relative to the directory of the scanned file, and this behavior should be documented at [https://pkg.go.dev/cmd/compile#hdr-Compiler_Directives](https://pkg.go.dev/cmd/compile#hdr-Compiler_Directives). This may require changes to the line directives emitted by cgo and other tools (including [unparam and unconvert](https://go.dev/cl/127658), whatever they are). Arguably, this is an incompatible change, but we notice that more often than not, when published Go modules use relative line directives, they already follow the proposed discipline. See also: - [https://github.com/golang/go/issues/24183](https://github.com/golang/go/issues/24183) - [https://github.com/golang/go/issues/26671](https://github.com/golang/go/issues/26671) - [https://github.com/golang/go/issues/69689](https://github.com/golang/go/issues/69689) @griesemer @ianlancetaylor @findleyr @timothy-king
Proposal
low
Critical
2,677,461,467
flutter
[Android] Dragging outside the text should not show the magnifier
### Steps to reproduce 1. Run an app with a text field (such as given below) on an Android device. 2. Enter some text. 3. Perform a drag inside the field but not on any of the text. ### Expected results The magnifier hides whenever the drag leaves the text. ### Actual results The magnifier is shown during the whole drag. ### Code sample <details><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); @override Widget build(BuildContext context) { return MaterialApp( title: 'Flutter Demo', theme: ThemeData( primarySwatch: Colors.blue, ), home: const MyHomePage(title: 'Flutter Demo Home Page'), ); } } class MyHomePage extends StatelessWidget { const MyHomePage({super.key, required this.title}); final String title; @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar( title: Text(title), ), body: Center( child: Column( mainAxisAlignment: MainAxisAlignment.center, children: const <Widget>[ TextField( ), ], ), ), ); } } ``` </details> ### Screenshots or Video <details open> <summary>Video of native Android</summary> https://github.com/user-attachments/assets/03f2f56a-25ec-47f7-be21-26b9037c4be1 </details> <details open> <summary>Video of Flutter</summary> https://github.com/user-attachments/assets/a295f2ee-5978-400a-a09f-97fd4760ebc6 </details> Discovered while working on https://github.com/flutter/flutter/pull/148784. ### Flutter Doctor output <details><summary>Doctor output</summary> ```console flutter doctor Building flutter tool... Resolving dependencies... (1.0s) Downloading packages... Got dependencies. Doctor summary (to see all details, run flutter doctor -v): [!] Flutter (Channel [user-branch], 3.22.0-35.0.pre.2441, on Debian GNU/Linux rodete 6.9.10-1rodete5-amd64, locale en_US.UTF-8) ! Flutter version 3.22.0-35.0.pre.2441 on channel [user-branch] at /usr/local/google/home/jmccandless/Projects/flutter Currently on an unknown channel. Run `flutter channel` to switch to an official channel. If that doesn't fix the issue, reinstall Flutter by following instructions at https://flutter.dev/setup. [โœ“] Android toolchain - develop for Android devices (Android SDK version 35.0.0) [โœ“] Chrome - develop for the web [โœ“] Linux toolchain - develop for Linux desktop [โœ“] Android Studio (version 2023.3) [โœ“] Connected device (3 available) [โœ“] Network resources ! Doctor found issues in 1 category. ``` </details>
platform-android,P2,team-text-input,triaged-text-input
low
Minor
2,677,551,234
opencv
Assertion error when flipping arrays with more than 512 channels
### System Information OpenCV python version: 4.10.0.84 Operating System / Platform: MacOS 15.1.1 Python version: 3.12.7 ### Detailed description Flipping arrays with more than 512 channels (third dimension) throws the following assertion error: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> cv2.error: OpenCV(4.10.0) /Users/xperience/GHA-Actions-OpenCV/_work/opencv-python/opencv-python/opencv/modules/core/src/matrix_transform.cpp:784: error: (-215:Assertion failed) _src.dims() <= 2 in function 'flip' ``` ### Steps to reproduce ```python import cv2 import numpy as np # Works arr1 = np.zeros((256, 256, 512), dtype=np.uint8) cv2.flip(arr1, 0) # Throws error: (-215:Assertion failed) arr2 = np.zeros((256, 256, 513), dtype=np.uint8) cv2.flip(arr2, 0) ``` ### Issue submission checklist - [X] I report the issue, it's not a question - [X] I checked the problem with documentation, FAQ, open issues, forum.opencv.org, Stack Overflow, etc and have not found any solution - [X] I updated to the latest OpenCV version and the issue is still there - [X] There is reproducer code and related data files (videos, images, onnx, etc)
bug,category: python bindings
low
Critical
2,677,551,526
rust
error: rustc interrupted by SIGSEGV
<!-- Thank you for filing a bug report! ๐Ÿ› Please provide a short summary of the bug, along with any information you feel relevant to replicating the bug. --> I tried this code: https://github.com/warcayac/axum_postgre_email ```rust cargo run ``` I expected to see this happen: *explanation* ``` โœ… Connection to the database is successful! ๐Ÿš€ Listening on 127.0.0.1:8080 ``` Instead, this happened: *explanation* ``` error: rustc interrupted by SIGSEGV, printing backtrace /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x33b6d83) [0x7a483d9b6d83] /usr/lib/libc.so.6(+0x3d1d0) [0x7a483a4171d0] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCs9f8yWUR3ijp_10rustc_lint7contextNtB4_9LintStore15check_lint_name+0x69b) [0x7a483f5cf0db] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4fd040d) [0x7a483f5d040d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9f8yWUR3ijp_10rustc_lint6levels22shallow_lint_levels_on+0x303) [0x7a483f8acb0d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x52ac7aa) [0x7a483f8ac7aa] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4c1d5) [0x7a483f14c1d5] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4be0d) [0x7a483f14be0d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCshbMUE98pR8y_12rustc_middle4lintNtB4_19ShallowLintLevelMap21lint_level_id_at_node+0x1757) [0x7a483f5dbb17] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor11visit_block+0x24b) [0x7a483f5d694b] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafety14check_unsafety+0x4ca) [0x7a483c19e84a] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513ff97) [0x7a483f73ff97] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x51401ca) [0x7a483f7401ca] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513fec1) [0x7a483f73fec1] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5449b0f) [0x7a483fa49b0f] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9EojWPpxxi0_15rustc_interface6passes8analysis+0x19) [0x7a483fa41da5] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5441d89) [0x7a483fa41d89] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6d22) [0x7a483fec6d22] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6acf) [0x7a483fec6acf] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x57860b2) [0x7a483fd860b2] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x583ed16) [0x7a483fe3ed16] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58779b0) [0x7a483fe779b0] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5877d2b) [0x7a483fe77d2b] /usr/lib/libc.so.6(+0x9439d) [0x7a483a46e39d] /usr/lib/libc.so.6(+0x11949c) [0x7a483a4f349c] note: we would appreciate a report at https://github.com/rust-lang/rust help: you can increase rustc's stack size by setting RUST_MIN_STACK=16777216 note: backtrace dumped due to SIGSEGV! resuming signal Finished `dev` profile [unoptimized + debuginfo] target(s) in 0.23s Running `target/debug/postgresql_email_backend` โœ… Connection to the database is successful! ๐Ÿš€ Listening on 127.0.0.1:8080 ``` After that I tried with: `RUST_MIN_STACK=16777216 cargo run`, same result ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> `rustc --version --verbose`: ``` rustc 1.82.0 (f6e511eec 2024-10-15) binary: rustc commit-hash: f6e511eec7342f59a25f7c0534f1dbea00d01b14 commit-date: 2024-10-15 host: x86_64-unknown-linux-gnu release: 1.82.0 LLVM version: 19.1.1 ``` <!-- Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your environment. E.g. `RUST_BACKTRACE=1 cargo build`. --> <details><summary>Backtrace</summary> <p> ``` RUST_BACKTRACE=1 cargo run error: rustc interrupted by SIGSEGV, printing backtrace /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x33b6d83) [0x7a483d9b6d83] /usr/lib/libc.so.6(+0x3d1d0) [0x7a483a4171d0] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCs9f8yWUR3ijp_10rustc_lint7contextNtB4_9LintStore15check_lint_name+0x69b) [0x7a483f5cf0db] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4fd040d) [0x7a483f5d040d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9f8yWUR3ijp_10rustc_lint6levels22shallow_lint_levels_on+0x303) [0x7a483f8acb0d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x52ac7aa) [0x7a483f8ac7aa] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4c1d5) [0x7a483f14c1d5] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4be0d) [0x7a483f14be0d] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCshbMUE98pR8y_12rustc_middle4lintNtB4_19ShallowLintLevelMap21lint_level_id_at_node+0x1757) [0x7a483f5dbb17] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor11visit_block+0x24b) [0x7a483f5d694b] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafety14check_unsafety+0x4ca) [0x7a483c19e84a] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513ff97) [0x7a483f73ff97] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x51401ca) [0x7a483f7401ca] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513fec1) [0x7a483f73fec1] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5449b0f) [0x7a483fa49b0f] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9EojWPpxxi0_15rustc_interface6passes8analysis+0x19) [0x7a483fa41da5] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5441d89) [0x7a483fa41d89] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6d22) [0x7a483fec6d22] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6acf) [0x7a483fec6acf] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x57860b2) [0x7a483fd860b2] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x583ed16) [0x7a483fe3ed16] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58779b0) [0x7a483fe779b0] /home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5877d2b) [0x7a483fe77d2b] /usr/lib/libc.so.6(+0x9439d) [0x7a483a46e39d] /usr/lib/libc.so.6(+0x11949c) [0x7a483a4f349c] note: we would appreciate a report at https://github.com/rust-lang/rust help: you can increase rustc's stack size by setting RUST_MIN_STACK=16777216 note: backtrace dumped due to SIGSEGV! resuming signal Finished `dev` profile [unoptimized + debuginfo] target(s) in 0.21s Running `target/debug/postgresql_email_backend` โœ… Connection to the database is successful! ๐Ÿš€ Listening on 127.0.0.1:8080 ^C ``` </p> </details>
I-crash,T-compiler,A-incr-comp,C-bug,S-needs-repro
low
Critical
2,677,577,476
pytorch
Dynamo graph break on RNN/LSTM doesn't resume the function correctly.
### ๐Ÿ› Describe the bug Dynamo purposely graph breaks on RNN, GRU, LSTMs. We expect the sub modules before and after the RNN/GRU/LSTM calls should be compiled as partial graphs, however, we found only the sub modules before the graph break got compiled, the one after doesn't. Repro: ``` import torch import torch.nn as nn import torch.nn.functional as F class MultiLayerModel(nn.Module): def __init__(self, input_dim, lstm_hidden_dim, conv_channels, linear_out_dim, num_classes): super(MultiLayerModel, self).__init__() # Convolutional layer self.conv = nn.Conv2d(in_channels=1, out_channels=conv_channels, kernel_size=3, stride=1, padding=1) # Layer normalization for convolution output self.conv_layer_norm = nn.LayerNorm([conv_channels, input_dim, input_dim]) # LSTM layer self.lstm = nn.LSTM(input_dim, lstm_hidden_dim, batch_first=True, bidirectional=True) # Fully connected linear layers self.fc1 = nn.Linear(lstm_hidden_dim * 2, linear_out_dim) # LSTM is bidirectional, so hidden_dim * 2 self.fc2 = nn.Linear(linear_out_dim, num_classes) # Layer normalization for linear layers self.linear_layer_norm = nn.LayerNorm(linear_out_dim) def forward(self, x): # Convolutional layer expects 4D input: (batch_size, channels, height, width) x = self.conv(x) x = F.relu(x) x = self.conv_layer_norm(x) # Flatten spatial dimensions for the LSTM (assuming square input images) batch_size, channels, height, width = x.size() x = x.view(batch_size, channels * height, width) # Prepare for LSTM: (batch, seq_len, features) # LSTM expects input: (batch_size, seq_len, input_dim) x, _ = self.lstm(x) # Fully connected layers x = F.relu(self.fc1(x[:, -1, :])) # Use the output from the last time step x = self.linear_layer_norm(x) x = self.fc2(x) return x input_dim = 28 lstm_hidden_dim = 128 conv_channels = 16 linear_out_dim = 64 num_classes = 10 model = MultiLayerModel(input_dim, lstm_hidden_dim, conv_channels, linear_out_dim, num_classes) print(model) model = torch.compile(model) # Dummy input: Batch of grayscale images (batch_size, channels, height, width) dummy_input = torch.randn(8, 1, 28, 28) output = model(dummy_input) print("Output shape:", output.shape) ``` Run this script with ```TORCH_LOGS=+dynamo```, you will find there is only one compiled graph which includes the sub module befroe line ```x = x.view(batch_size, channels * height, width)```. ### Versions N/A cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @amjames
triaged,bug,oncall: pt2,module: dynamo
low
Critical
2,677,578,157
go
build: build failure on go1.23-linux-amd64_c3h88-perf_vs_release
``` #!watchflakes default <- builder == "go1.23-linux-amd64_c3h88-perf_vs_release" && repo == "go" && mode == "build" ``` Issue created automatically to collect these failures. Example ([log](https://ci.chromium.org/b/8730818689678245169)): go: downloading github.com/BurntSushi/toml v1.0.0 go: downloading github.com/kballard/go-shellquote v0.0.0-20180428030007-95032a82bc51 2024/11/19 18:09:12 Load average: 0.67 0.20 0.07 1/1176 1834 2024/11/19 18:09:12 Waiting for load average to drop below 0.20... 2024/11/19 18:09:42 Load average: 0.47 0.19 0.07 1/1177 1834 2024/11/19 18:09:42 Waiting for load average to drop below 0.20... 2024/11/19 18:10:12 Load average: 0.28 0.17 0.07 2/1177 1834 2024/11/19 18:10:12 Waiting for load average to drop below 0.20... 2024/11/19 18:10:42 Load average: 0.23 0.17 0.07 3/1180 1834 2024/11/19 18:10:42 Waiting for load average to drop below 0.20... ... [sweet] Running benchmark tile38 for experiment: run 8 [sweet] Running benchmark tile38 for baseline: run 8 [sweet] Running benchmark tile38 for experiment: run 9 [sweet] Running benchmark tile38 for baseline: run 9 [sweet] Running benchmark tile38 for experiment: run 10 [sweet] Running benchmark tile38 for baseline: run 10 [sweet] error: failed to execute benchmarks: gvisor 2024/11/19 23:45:15 Error running sweet: error running sweet run: exit status 1 2024/11/19 23:45:15 FAIL exit status 1 โ€” [watchflakes](https://go.dev/wiki/Watchflakes)
NeedsInvestigation
low
Critical