id
int64
393k
2.82B
repo
stringclasses
68 values
title
stringlengths
1
936
body
stringlengths
0
256k
labels
stringlengths
2
508
priority
stringclasses
3 values
severity
stringclasses
3 values
2,805,711,342
vscode
"NVDA Not Reading When I Open a Folder from Within VSCode"
Type: <b>Bug</b> When I open a folder within Visual Studio Code, NVDA announces the folder as "unknown" instead of reading anything . i have to restart nvda and alt tab out and back in. This issue prevents me from knowing which folder or files are being opened, making it difficult to navigate effectively within VSCode. note, i am on a mac with a windows VM. I was also able to reproduce on my other computers. VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Windows_NT arm64 10.0.26100 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple Silicon (4 x 3200)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: unavailable_off<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|undefined| |Memory (System)|7.99GB (0.85GB free)| |Process Argv|--crash-reporter-id c37eff74-c330-4ccb-910b-e92d95621bcc| |Screen Reader|yes| |VM|100%| </details><details><summary>Extensions (4)</summary> Extension|Author (truncated)|Version ---|---|--- remote-ssh|ms-|0.116.1 remote-ssh-edit|ms-|0.87.0 remote-explorer|ms-|0.4.3 pythagora-vs-code|pyt|1.2.13 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vspor879:30202332 vspor708:30202333 vspor363:30204092 vscod805:30301674 binariesv615:30325510 vsaa593:30376534 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 2e7ec940:31000449 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupytercf:31046870 2f103344:31071589 nativerepl1:31139838 pythonrstrctxt:31112756 nativeloc1:31192215 cf971741:31144450 iacca1:31171482 notype1:31157159 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 6074i472:31201624 dwoutputs:31217127 hdaa2157:31222309 copilot_t_ci:31222730 ``` </details> <!-- generated by issue reporter -->
info-needed
low
Critical
2,805,719,003
react
Bug: Synchronously Restart Render on Error During Async Rendering to Prevent Inconsistencies
Synchronously restart when an error is thrown during async rendering ### Description In async mode, events are interleaved with rendering. If an event mutates state that is later accessed during render, it can cause inconsistencies. Restarting the render synchronously resolves the inconsistency and prevents further mutations during interleaved events. ### Steps to Reproduce 1. Enable async rendering mode. 2. Trigger an event that mutates state during rendering. 3. Observe inconsistencies during the render process. ### Expected Behavior The render should restart synchronously to prevent inconsistencies and further mutations from interfering. ### Actual Behavior Inconsistent rendering occurs, leading to potential UI issues or errors. ### Suggested Fix Implement synchronous restart of the render when an error is thrown during async rendering to handle the interleaved events and ensure consistency.
Status: Unconfirmed,Resolution: Needs More Information
medium
Critical
2,805,742,425
angular
Expose `createInputSignal`, or ease restrictions on static analysis for Signal functions to reduce boilerplate
### Which @angular/* package(s) are relevant/related to the feature request? core ### Description The Angular CDK provides `coercion` functions to make it easy to transform `input` values. For example: ```ts import { BooleanInput, coerceBooleanProperty } from '@angular/cdk/coercion'; public readonly canEdit = input<boolean, BooleanInput>(false, { transform: coerceBooleanProperty }); public readonly canDelete = input<boolean, BooleanInput>(false, { transform: coerceBooleanProperty }); ``` This makes it easy to use the component like below, where simply the presence of an empty `canEdit` or `canDelete` attribute will set the `input` Signal's value to `true`: ```html <my-component canEdit canDelete /> ``` When using something like Prettier with an 80-char line length, each input defined like this becomes three lines long: ```ts public readonly canEdit = input(false, { transform: coerceBooleanProperty, }); public readonly canDelete = input(true, { transform: coerceBooleanProperty, }); ``` This is a lot of boilerplate. Ideally, I'd be able to do create a utility function like the following: ```ts export function booleanInput(defaultValue: boolean): InputSignalWithTransform<boolean, BooleanInput> { return input(defaultValue, { transform: coerceBooleanProperty }); } ``` ```ts public readonly canEdit = booleanInput(false); public readonly canDelete = booleanInput(false); ``` However the above code isn't valid, because: `✘ [ERROR] TS-998110: Unsupported call to the input function. This function can only be called in the initializer of a class member. [plugin angular-compiler]` That's fair, so next I figured I'd try and simplify things by doing: ```ts const boolTransform: InputOptionsWithTransform<boolean, BooleanInput> = { transform: coerceBooleanProperty }; ``` ```ts public readonly canEdit = input(false, boolTransform); public readonly canDelete = input(false, boolTransform); ``` However this **also** isn't valid, because: `✘ [ERROR] TS-991010: Argument needs to be an object literal that is statically analyzable. [plugin angular-compiler]` I understand why both of these restrictions exist, but it would be nice to be able to avoid needing to rewrite out the entire options object for every Signal when they are all exactly the same. ### Proposed solution I'm unsure of any knock-on effects of exposing it, but `core/src/authoring/input` contains the below function: ```ts export function createInputSignal<T, TransformT>( initialValue: T, options?: InputOptions<T, TransformT>, ): InputSignalWithTransform<T, TransformT>; ``` If this were exposed from `@angular/core`, it could be possible to write wrapper utilities for signals such as the `booleanInput()` example I described, or other shortcuts such as a number-based signal that has min/max values and normalizes values to be within range, etc. ### Alternatives considered Alternatively, relaxing (or expanding) the static analysis of Options objects would mean that it would be easier to repeat the same transforms/options throughout the codebase even when using the base `input`/`model`/etc functions.
area: compiler,cross-cutting: signals
low
Critical
2,805,760,863
ollama
API parameter: 'reasoning_effort' (for DeepSeek-R1)
When I will be able to use `reasoning_effort` like on DeepSeek-R1? :) Doc: https://api-docs.deepseek.com/guides/reasoning_model
feature request
low
Minor
2,805,761,049
react
Bug: Prevent event log from growing unbounded
When a Scheduler profile runs without stopping, the event log will grow unbounded. Eventually it will run out of memory and the VM will throw an error. React version: [v16.9.0](https://github.com/facebook/react/releases/tag/v16.9.0) ## Steps To Reproduce 1. Run Scheduler profile 2. Watch memory resource growing ## The current behavior When Scheduler profile starts it runs out of memory and the JVM throws an error ## The expected behavior The event log should not grow unbounded
Status: Unconfirmed
medium
Critical
2,805,765,889
flutter
[go_router] [go_router_builder] go_router already supports preloading but go_router_builder doesn't
### Use case Support preload for `StatefulShellBranch` when use `go_router_builder` ### Proposal Add `preload` field to `go_router_builder` and code gen helper function of `go_router`
c: new feature,package,c: proposal,p: go_router_builder,team-go_router
low
Minor
2,805,773,579
pytorch
[Dynamo]while_loop raise an exception
### 🐛 Describe the bug while loop will raise an exception. The reason is input tensor has been sliced. ``` tensor strided info: shape: torch.Size([192]), original_tshape: [64], slice:slice(None, None, 3) ``` This creates tensor with stride=3, but once the body_fn execute tensor+1 the stride is reset to 1 and this causes the error. If tensor which is strided by nature due to slice before passing to while_loop then pytorch doesn't accept this. mini reproducer: ``` import torch class OpWrapperModule(torch.nn.Module): def __init__(self): super().__init__() def forward(self, ifm, op_inputs_dict): result = torch.while_loop(**op_inputs_dict) return result torch.manual_seed(7616) ifm_t = torch.randn([64]) ifm = ifm_t[slice(None, None, 3)] iterations = torch.tensor(50) def cond_fn(inputs, iterations): return iterations > 0 def body_fn(inputs, iterations): return [inputs + 1, iterations - 1] params = { "cond_fn": cond_fn, "body_fn": body_fn, "carried_inputs": (ifm, iterations), } model = OpWrapperModule() result = model(ifm, params) ``` ERROR log and trace: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 55, in graph_break_as_hard_error return fn(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 1115, in call_function unimplemented( File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/exc.py", line 317, in unimplemented raise Unsupported(msg, case_name=case_name) torch._dynamo.exc.Unsupported: Expected carried_inputs and body outputs return tensors with same metadata but find: pair[0] differ in 'stride: (3,) vs (1,)', where lhs is TensorMetadata(shape=torch.Size([22]), dtype=torch.float32, requires_grad=False, stride=(3,), memory_format=None, is_quantized=False, qparams={}) and rhs is TensorMetadata(shape=torch.Size([22]), dtype=torch.float32, requires_grad=False, stride=(1,), memory_format=None, is_quantized=False, qparams={}) The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/zhenzhao/qnpu/dynamo/src/pytorch-training-tests/tests/torch_feature_val/single_op/repro_while_loop.py", line 28, in <module> result = model(ifm, params) File "/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py", line 1742, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py", line 1753, in _call_impl return forward_call(*args, **kwargs) File "/home/zhenzhao/qnpu/dynamo/src/pytorch-training-tests/tests/torch_feature_val/single_op/repro_while_loop.py", line 8, in forward result = torch.while_loop(**op_inputs_dict) File "/usr/local/lib/python3.10/dist-packages/torch/_higher_order_ops/while_loop.py", line 165, in while_loop return torch.compile( File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/eval_frame.py", line 574, in _fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 1380, in __call__ return self._torchdynamo_orig_callable( File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 547, in __call__ return _compile( File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 986, in _compile guarded_code = compile_inner(code, one_graph, hooks, transform) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 715, in compile_inner return _compile_inner(code, one_graph, hooks, transform) File "/usr/local/lib/python3.10/dist-packages/torch/_utils_internal.py", line 95, in wrapper_function return function(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 750, in _compile_inner out_code = transform_code_object(code, transform) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/bytecode_transformation.py", line 1361, in transform_code_object transformations(instructions, code_options) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 231, in _fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 662, in transform tracer.run() File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 2868, in run super().run() File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1052, in run while self.step(): File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 962, in step self.dispatch_table[inst.opcode](self, inst) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 659, in wrapper return inner_fn(self, inst) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1736, in CALL_FUNCTION_EX self.call_function(fn, argsvars.items, kwargsvars) File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 897, in call_function self.push(fn.call_function(self, args, kwargs)) # type: ignore[arg-type] File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 58, in graph_break_as_hard_error raise UncapturedHigherOrderOpError(reason + msg) from e torch._dynamo.exc.UncapturedHigherOrderOpError: while_loop doesn't work unless it is captured completely with torch.compile. Scroll up to find out what causes the graph break. from user code: File "/usr/local/lib/python3.10/dist-packages/torch/_higher_order_ops/while_loop.py", line 156, in _while_loop_op_wrapper return while_loop_op(*args, **kwargs) ### Versions PyTorch version: 2.6.0a0+git30ac7fd Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: Ubuntu 22.04 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: 14.0.5 (ssh://[[email protected]](mailto:[email protected])/habana-internal/tpc_llvm10 150d2d7c6a8ff8abf0d8ce194d3fac3986b078e6) CMake version: version 3.28.1 Libc version: glibc-2.35 Python version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-5.15.0-127-generic-x86_64-with-glibc2.35 Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 43 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 12 On-line CPU(s) list: 0-11 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Gold 6132 CPU @ 2.60GHz CPU family: 6 Model: 85 Thread(s) per core: 1 Core(s) per socket: 6 Socket(s): 2 Stepping: 0 BogoMIPS: 5187.81 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon nopl xtopology tsc_reliable nonstop_tsc cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch invpcid_single pti ssbd ibrs ibpb stibp tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 invpcid avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xsaves arat pku ospke md_clear flush_l1d arch_capabilities Virtualization: VT-x Hypervisor vendor: VMware Virtualization type: full L1d cache: 384 KiB (12 instances) L1i cache: 384 KiB (12 instances) L2 cache: 12 MiB (12 instances) L3 cache: 38.5 MiB (2 instances) NUMA node(s): 1 NUMA node0 CPU(s): 0-11 Vulnerability Gather data sampling: Unknown: Dependent on hypervisor status Vulnerability Itlb multihit: KVM: Mitigation: VMX disabled Vulnerability L1tf: Mitigation; PTE Inversion; VMX flush not necessary, SMT disabled Vulnerability Mds: Mitigation; Clear CPU buffers; SMT Host state unknown Vulnerability Meltdown: Mitigation; PTI Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT Host state unknown Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Mitigation; IBRS Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI SW loop, KVM SW loop Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected cc @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225
triaged,oncall: pt2,module: higher order operators,module: pt2-dispatcher
low
Critical
2,805,777,333
PowerToys
Keeps forgetting that I have fancy zones turned on.
### Microsoft PowerToys version 0.87.1 ### Installation method Other (please specify in "Steps to Reproduce") ### Running as admin Yes ### Area(s) with issue? FancyZones ### Steps to reproduce Installation method don't remember. From Google search. It randomly forgets that I have fancy zones turned on (great feature, b.t.w. :) ), and I have to manually turn it on again. ### ✔️ Expected Behavior It remembers that I have fancy zones turned on, for each user. ### ❌ Actual Behavior It randomly forgets. ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,805,797,414
react
Bug: Fix Buffer - Write Chunks to Separate Buffers
<!-- Please provide a clear and concise description of what the bug is. Include screenshots if needed. Please test using the latest version of the relevant React packages to make sure your issue has not already been fixed. --> The bug happens when using ReactDOMFizzServer.renderToReadableStream to render big chunks of data. If a chunk is bigger than the buffer (default: 512 bytes), the buffer overflows or uses incomplete data. This causes wrong or cut-off output because the renderer doesn’t handle buffer size properly. React version: React 19.0.0 ## Steps To Reproduce 1. Create a large chunk of data that is bigger than the buffer size. 2. Start processing the data using `ReactDOMFizzServer.renderToReadableStream`. 3. Observe the behavior when the chunks exceed the buffer size. <!-- Your bug will get fixed much faster if we can run your code and it doesn't have dependencies other than React. Issues without reproduction steps or code examples may be immediately closed as not actionable. --> <!-- Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a repository on GitHub, or provide a minimal code example that reproduces the problem. You may provide a screenshot of the application if you think it is relevant to your bug report. Here are some tips for providing a minimal example: https://stackoverflow.com/help/mcve. --> ``` import ReactDOMFizzServer from 'react-dom/server'; // Minimal data to reproduce the problem const largeChunk = 'A'.repeat(2049); // A chunk exceeding default 512 bytes async function testStream() { // simple React structure with the large chunk const stream = await ReactDOMFizzServer.renderToReadableStream( <div>{largeChunk}</div> ); // function to read the stream const decoder = new TextDecoder(); const reader = stream.getReader(); let result = ''; while (true) { const { done, value } = await reader.read(); if (done) break; result += decoder.decode(value); } console.log(result); } testStream(); ``` ## The current behavior Right now, when chunks are sent to a `ReadableStream`, they are enqueued directly without checking the buffer size. If a chunk is too big (like 2049 bytes, while the buffer is only 512 bytes), it reuses old buffers, which messes up the output. For example, instead of properly rendering, we get something like ```<div>AAAAA...</div>```. ## The expected behavior Chunks should be written to their own buffer and split into smaller parts if they are too big. The buffer should only be sent once it's full or the writing is complete. This way, even large chunks (like 2049 bytes) will render correctly, for example: ```<div>AAAAAAAAAAAAAAAAA...</div>```.
Status: Unconfirmed
medium
Critical
2,805,814,674
pytorch
[CUDA] Illegal Memory Access with `AdaptiveMaxPool2d`
### 🐛 Describe the bug Found by fuzzer, the parameter used to trigger this does not look too corner-case. @eqy ```python import torch m1 = torch.randn(8812, 1, 2).cuda() model = torch.nn.AdaptiveMaxPool2d(output_size=[262143, 1]).cuda() model(m1) ``` ```bash computer-sanitizer python3 poc7.py ``` compute-sanitizer log. ```python ========= Invalid __global__ write of size 8 bytes ========= at void at::native::<unnamed>::adaptivemaxpool<float>(const T1 *, T1 *, long *, int, int, int, int, long, long, long)+0x1b80 ========= by thread (0,4,0) in block (8195,0,0) ========= Address 0x79d8025f0008 is out of bounds ========= and is 17,173,643,256 bytes before the nearest allocation at 0x79dc02000000 of size 18,480,103,424 bytes ========= Saved host backtrace up to driver entry point at kernel launch time ========= Host Frame: [0x2dfbef] ========= in /lib/x86_64-linux-gnu/libcuda.so.1 ========= Host Frame: [0x15803] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/../../../../libcudart.so.12 ========= Host Frame:cudaLaunchKernel [0x75230] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/../../../../libcudart.so.12 ========= Host Frame:at::native::structured_adaptive_max_pool2d_out_cuda::impl(at::Tensor const&, c10::ArrayRef<long>, at::Tensor const&, at::Tensor const&)::{lambda()#1}::operator()() const [0x160853b] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so ========= Host Frame:at::native::structured_adaptive_max_pool2d_out_cuda::impl(at::Tensor const&, c10::ArrayRef<long>, at::Tensor const&, at::Tensor const&) [0x160a8f4] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so ========= Host Frame:at::(anonymous namespace)::wrapper_CUDA_adaptive_max_pool2d(at::Tensor const&, c10::ArrayRef<long>) [0x3606f1e] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so ========= Host Frame:c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<std::tuple<at::Tensor, at::Tensor> (at::Tensor const&, c10::ArrayRef<long>), &at::(anonymous namespace)::wrapper_CUDA_adaptive_max_p ool2d>, std::tuple<at::Tensor, at::Tensor>, c10::guts::typelist::typelist<at::Tensor const&, c10::ArrayRef<long> > >, std::tuple<at::Tensor, at::Tensor> (at::Tensor const&, c10::ArrayRef<long>)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, c10::Array Ref<long>) [0x3606fe2] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so ========= Host Frame:at::_ops::adaptive_max_pool2d::redispatch(c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x27406bd] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so ========= Host Frame:torch::autograd::VariableType::(anonymous namespace)::adaptive_max_pool2d(c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x49ee7d4] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so ========= Host Frame:c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<std::tuple<at::Tensor, at::Tensor> (c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>), &torch::autograd::VariableType::(a nonymous namespace)::adaptive_max_pool2d>, std::tuple<at::Tensor, at::Tensor>, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long> > >, std::tuple<at::Tensor, at::Tensor> (c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>)>:: call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x49eef95] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so ========= Host Frame:at::_ops::adaptive_max_pool2d::call(at::Tensor const&, c10::ArrayRef<long>) [0x27cb9db] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so ========= Host Frame:torch::autograd::THPVariable_adaptive_max_pool2d(_object*, _object*, _object*) [0x765397] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_python.so ========= Host Frame:cfunction_call in /usr/local/src/conda/python-3.12.7/Objects/methodobject.c:537 [0x149d53] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyObject_MakeTpCall in /usr/local/src/conda/python-3.12.7/Objects/call.c:240 [0x11af9a] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:2715 [0x125902] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:method_vectorcall in /usr/local/src/conda/python-3.12.7/Objects/classobject.c:91 [0x175716] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:3263 [0x12acac] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:method_vectorcall in /usr/local/src/conda/python-3.12.7/Objects/classobject.c:91 [0x175716] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:3263 [0x12acac] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyObject_FastCallDictTstate in /usr/local/src/conda/python-3.12.7/Objects/call.c:133 [0x11db06] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyObject_Call_Prepend in /usr/local/src/conda/python-3.12.7/Objects/call.c:508 [0x157e55] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:slot_tp_call in /usr/local/src/conda/python-3.12.7/Objects/typeobject.c:8782 [0x22f065] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyObject_MakeTpCall in /usr/local/src/conda/python-3.12.7/Objects/call.c:240 [0x11af9a] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:2715 [0x125902] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:PyEval_EvalCode in /usr/local/src/conda/python-3.12.7/Python/ceval.c:578 [0x1e3c6d] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:run_eval_code_obj in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1722 [0x20a0b6] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:run_mod in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1743 [0x2056d6] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:pyrun_file in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1643 [0x21d601] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyRun_SimpleFileObject in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:433 [0x21cf3f] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:_PyRun_AnyFileObject in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:78 [0x21cd32] ========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3 ========= Host Frame:Py_RunMain in /usr/local/src/conda/python-3.12.7/Modules/main.c:713 [0x215dc2] [ ``` ### Versions ``` Collecting environment information... PyTorch version: 2.5.1 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 24.04.1 LTS (x86_64) GCC version: (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version: Could not collect CMake version: Could not collect Libc version: glibc-2.39 Python version: 3.12.7 | packaged by Anaconda, Inc. | (main, Oct 4 2024, 13:27:36) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-6.11.0-1007-oem-x86_64-with-glibc2.39 Is CUDA available: True CUDA runtime version: 12.6.85 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA H100 PCIe Nvidia driver version: 560.35.05 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 48 On-line CPU(s) list: 0-47 Vendor ID: GenuineIntel Model name: INTEL(R) XEON(R) SILVER 4510 CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 12 Socket(s): 2 Stepping: 8 CPU(s) scaling MHz: 37% CPU max MHz: 4100.0000 CPU min MHz: 800.0000 BogoMIPS: 4800.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust sgx bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd sgx_lc fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 1.1 MiB (24 instances) L1i cache: 768 KiB (24 instances) L2 cache: 48 MiB (24 instances) L3 cache: 60 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-11,24-35 NUMA node1 CPU(s): 12-23,36-47 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI BHI_DIS_S Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] numpy==2.1.3 [pip3] nvidia-cublas-cu12==12.6.4.1 [pip3] nvidia-cudnn-cu12==9.5.1.17 [pip3] torch==2.5.1 [pip3] torchaudio==2.5.1 [pip3] torchvision==0.20.1 [pip3] triton==3.1.0 [conda] blas 1.0 mkl [conda] cuda-cudart 12.4.127 0 nvidia [conda] cuda-cupti 12.4.127 0 nvidia [conda] cuda-libraries 12.4.1 0 nvidia [conda] cuda-nvrtc 12.4.127 0 nvidia [conda] cuda-nvtx 12.4.127 0 nvidia [conda] cuda-opencl 12.6.77 0 nvidia [conda] cuda-runtime 12.4.1 0 nvidia [conda] ffmpeg 4.3 hf484d3e_0 pytorch [conda] libcublas 12.4.5.8 0 nvidia [conda] libcufft 11.2.1.3 0 nvidia [conda] libcurand 10.3.7.77 0 nvidia [conda] libcusolver 11.6.1.9 0 nvidia [conda] libcusparse 12.3.1.170 0 nvidia [conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch [conda] libnvjitlink 12.4.127 0 nvidia [conda] mkl 2023.1.0 h213fc3f_46344 [conda] mkl-service 2.4.0 py312h5eee18b_1 [conda] mkl_fft 1.3.11 py312h5eee18b_0 [conda] mkl_random 1.2.8 py312h526ad5a_0 [conda] numpy 2.1.3 py312hc5e2394_0 [conda] numpy-base 2.1.3 py312h0da6c21_0 [conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi [conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi [conda] pytorch 2.5.1 py3.12_cuda12.4_cudnn9.1.0_0 pytorch [conda] pytorch-cuda 12.4 hc786d27_7 pytorch [conda] pytorch-mutex 1.0 cuda pytorch [conda] torchaudio 2.5.1 py312_cu124 pytorch [conda] torchtriton 3.1.0 py312 pytorch [conda] torchvision 0.20.1 py312_cu124 pytorch ``` cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @ptrblck @msaroufim @eqy
module: nn,module: cuda,triaged,module: edge cases
low
Critical
2,805,822,844
pytorch
torch._dynamo.exc.Unsupported: Graph break due to unsupported builtin torch._C._dynamo.eval_frame.set_eval_frame.
### 🐛 Describe the bug Part of an example as the following: ```python from torch_mlir import fx module = fx.export_and_import( model, input_tensor, output_type="tosa", func_name="forward", ) ``` when I use pyinstaller to pack the fx graph export and import to a binary file, it encounters the following problem. Actually, when I run the unpacked python code, it works well. ![Image](https://github.com/user-attachments/assets/d0dc84cb-9ae8-4c25-afe5-0cf43e41667d) ### Versions `torch==2.6.0.dev20241216+cpu` cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4
oncall: pt2,export-triaged,oncall: export
low
Critical
2,805,827,334
rust
Error installing cargo-watch
<!-- Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for how to create smaller examples. http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/ --> ### Code ```Rust cargo install cargo-watch ``` ### Meta `rustc --version --verbose`: ``` rustc 1.84.0 (9fc6b4312 2025-01-07) binary: rustc commit-hash: 9fc6b43126469e3858e2fe86cafb4f0fd5068869 commit-date: 2025-01-07 host: x86_64-unknown-linux-gnu release: 1.84.0 LLVM version: 19.1.5 ``` ### Error output ``` error: the compiler unexpectedly panicked. this is a bug. note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md note: rustc 1.84.0 (9fc6b4312 2025-01-07) running on x86_64-unknown-linux-gnu note: compiler flags: --crate-type lib -C opt-level=3 -C panic=abort -C linker-plugin-lto -C codegen-units=1 -C strip=debuginfo note: some of the compiler flags provided by cargo are hidden ``` ``` thread 'main' panicked at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/thread/mod.rs:1719:40: called `Option::unwrap()` on a `None` value stack backtrace: Compiling inotify v0.7.1 0: 0x7148abd8682a - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::ha4a311b32f6b4ad8 1: 0x7148ac4277e2 - core::fmt::write::h1866771663f62b81 2: 0x7148ad281d51 - std::io::Write::write_fmt::hb549e7444823135e 3: 0x7148abd86682 - std::sys::backtrace::BacktraceLock::print::hddd3a9918ce29aa7 4: 0x7148abd88b5a - std::panicking::default_hook::{{closure}}::h791f75256b902d7d 5: 0x7148abd889c0 - std::panicking::default_hook::h82cc572fcb0d8cd7 6: 0x7148aae14f55 - std[1b49f43dde054edc]::panicking::update_hook::<alloc[f0e0d4128a1437e6]::boxed::Box<rustc_driver_impl[c421ed190efad9be]::install_ice_hook::{closure#0}>>::{closure#0} 7: 0x7148abd89238 - std::panicking::rust_panic_with_hook::he21644cc2707f2c4 8: 0x7148abd88fd6 - std::panicking::begin_panic_handler::{{closure}}::h42f7c414fed3cad9 9: 0x7148abd86cd9 - std::sys::backtrace::__rust_end_short_backtrace::ha26cf5766b4e8c65 10: 0x7148abd88ccc - rust_begin_unwind 11: 0x7148a88e43f0 - core::panicking::panic_fmt::h74866b78e934b1c0 12: 0x7148a8afe56c - core::panicking::panic::h95d8269cf8bd4f7a 13: 0x7148a9e65909 - core::option::unwrap_failed::hca433a9693b91bd2 14: 0x7148ad27a0d9 - rustc_driver_impl[c421ed190efad9be]::run_compiler 15: 0x7148ad1c99e0 - rustc_driver_impl[c421ed190efad9be]::main 16: 0x5eadf5e9e497 - rustc_main[302bc4b73869fe48]::main 17: 0x5eadf5e9e483 - std[1b49f43dde054edc]::sys::backtrace::__rust_begin_short_backtrace::<fn(), ()> 18: 0x5eadf5e9e479 - std[1b49f43dde054edc]::rt::lang_start::<()>::{closure#0} 19: 0x7148ad29dd41 - std::rt::lang_start_internal::h78dd36c15a6b42b8 20: 0x5eadf5ea5ea7 - main 21: 0x7148a7429d90 - <unknown> 22: 0x7148a7429e40 - __libc_start_main 23: 0x5eadf5ea5db6 - <unknown> 24: 0x0 - <unknown> Compiling clap v2.34.0 ``` </details>
I-ICE,T-compiler,C-bug,S-needs-repro
low
Critical
2,805,829,310
pytorch
DISABLED test_tensor_subclass_basic (__main__.TestCompiledAutograd)
Platforms: asan, linux, rocm, mac, macos, slow This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_tensor_subclass_basic&suite=TestCompiledAutograd&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/36028557832). Over the past 3 hours, it has been determined flaky in 20 workflow(s) with 40 failures and 20 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_tensor_subclass_basic` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `inductor/test_compiled_autograd.py` cc @clee2000 @wdvr @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @chauhang @aakhundov
triaged,module: flaky-tests,skipped,oncall: pt2,module: inductor
low
Critical
2,805,848,501
electron
I am currently upgrading the project from Electron 15 to 29, but the embedded document websites consistently pop up a font download popup window
### Preflight Checklist - [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success. ### Electron Version 29.4.26 ### What operating system(s) are you using? macOS ### Operating System Version 13.2.1 (22D68) ### What arch are you using? arm64 (including Apple Silicon) ### Last Known Working Electron version 30.0.0 ### Expected Behavior I hope that, just like in Electron 15, this pop - up can be skipped entirely, and the default font can be used directly. This is because such an experience is really bad for users. ### Actual Behavior Request for Support Background: I am currently upgrading the project from Electron 15 to Electron 29, but the embedded document websites consistently pop up a font download popup window. Problem: When using BrowserView, WebView, or iframe in the project to reference external websites, this popup window will appear when a specific font does not exist on Mac. This font download popup window did not exist in version 15. I want to keep it consistent with version 15. Which feature of Electron or Chrome has led to this phenomenon? How should I set it? Below is a screenshot. Additionally, through my tests in fiddle, I found that this issue still persists in the last version 29.4.26 of Electron 29. However, this problem does not occur in version 30.0.0. Also, there is no such problem in the last version of Electron 28 ![Image](https://github.com/user-attachments/assets/a4568b5c-413b-42a2-936d-aefd99e539a8) ### Testcase Gist URL https://gist.github.com/Tsite2007/aee74ff4afe6f14148d88a1aacb74364 ### Additional Information _No response_
platform/macOS,blocked/need-info ❌,bug :beetle:,has-repro-gist
low
Critical
2,805,853,403
flutter
Engine `local_engine_builds` failing since #161825
`local_engine_builds` has been in bringup, which means https://github.com/flutter/flutter/pull/161825 broke it without knowing. (https://github.com/flutter/flutter/pull/160627 initially marked it as `bringup: true`, @jonahwilliams was this intended to be permanent?) @eyebrowsoffire can you fix this (or revert your changes if you don't imagine this happening for some time)?
engine,P1,c: tech-debt,team-engine,triaged-engine,fyi-web,e: engine-tool
medium
Minor
2,805,876,375
vscode
Terminal font ligatures doesn't render as expected
In terminal ![Image](https://github.com/user-attachments/assets/2f91f1d0-3c7f-4d73-ae0e-c8aa7d805f2d) In editor ![Image](https://github.com/user-attachments/assets/ff35a66e-7c3e-42fe-b576-997f38d3aca5) Version: 1.97.0-insider (user setup) Commit: d226a2a497b928d78aa654f74c8af5317d3becfb Date: 2025-01-22T05:05:14.565Z Electron: 32.2.7 ElectronBuildId: 10660205 Chromium: 128.0.6613.186 Node.js: 20.18.1 V8: 12.8.374.38-electron.0 OS: Windows_NT x64 10.0.22621
info-needed,terminal-ligatures
low
Major
2,805,889,854
pytorch
Flex Attention not support score_mod with gradients
### 🐛 Describe the bug `Flex Attention` does not support `score_mod with gradients`, making it impossible to define a learnable score_mod for the dynamic mask attention variants. ```python import torch from torch.nn.attention.flex_attention import flex_attention A = torch.nn.Parameter(torch.zeros(1)) query = torch.randn(1, 1, 1, 32) key = torch.randn(1, 1, 1, 32) value = torch.randn(1, 1, 1, 32) dt = torch.randn(1, 1, 1) dynamic_mask = torch.exp(A * dt).transpose(-1, -2) attn_mask = dynamic_mask[:, :, None, :] def dynamic_mod(score, batch, head, q_idx, kv_idx): score = score + attn_mask[batch][head][0][kv_idx] return score attn_output = flex_attention( query=query, key=key, value=value, score_mod=dynamic_mod, ) print(attn_output) ``` ``` Traceback (most recent call last): File "e:\Doge\bug.py", line 22, in <module> attn_output = flex_attention( File "E:\conda\envs\doge\lib\site-packages\torch\nn\attention\flex_attention.py", line 1038, in flex_attention out, lse = torch.compile( File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn return fn(*args, **kwargs) File "E:\conda\envs\doge\lib\site-packages\torch\nn\attention\flex_attention.py", line 1032, in _flex_attention_hop_wrapper def _flex_attention_hop_wrapper(*args, **kwargs): File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn return fn(*args, **kwargs) File "<eval_with_key>.3", line 24, in forward File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 109, in __call__ return super().__call__( File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 433, in __call__ return wrapper() File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn return fn(*args, **kwargs) File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 429, in wrapper return self.dispatch( File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 412, in dispatch return kernel(*args, **kwargs) File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 703, in flex_attention_autograd out, logsumexp = FlexAttentionAutogradOp.apply( File "E:\conda\envs\doge\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 578, in forward assert ( AssertionError: Captured buffers that require grad are not yet supported. ``` ### Versions Collecting environment information... PyTorch version: 2.6.0a0+df5bbc09d1.nv24.12 Is debug build: False CUDA used to build PyTorch: 12.6 ROCM used to build PyTorch: N/A OS: Ubuntu 24.04.1 LTS (x86_64) GCC version: (Ubuntu 13.2.0-23ubuntu4) 13.2.0 Clang version: Could not collect CMake version: version 3.31.1 Libc version: glibc-2.39 Python version: 3.12.3 (main, Nov 6 2024, 18:32:19) [GCC 13.2.0] (64-bit runtime) Python platform: Linux-5.15.153.1-microsoft-standard-WSL2-x86_64-with-glibc2.39 Is CUDA available: True CUDA runtime version: 12.6.85 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA GeForce RTX 4090 Nvidia driver version: 566.36 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.6.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.6.0 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Vendor ID: GenuineIntel Model name: 13th Gen Intel(R) Core(TM) i9-13900KS CPU family: 6 Model: 183 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 1 Stepping: 1 BogoMIPS: 6374.42 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni umip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm md_clear serialize flush_l1d arch_capabilities Virtualization: VT-x Hypervisor vendor: Microsoft Virtualization type: full L1d cache: 768 KiB (16 instances) L1i cache: 512 KiB (16 instances) L2 cache: 32 MiB (16 instances) L3 cache: 36 MiB (1 instance) Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Retbleed: Mitigation; Enhanced IBRS Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.26.4 [pip3] nvidia-cudnn-frontend==1.8.0 [pip3] nvtx==0.2.5 [pip3] onnx==1.17.0 [pip3] optree==0.13.1 [pip3] pynvjitlink==0.3.0 [pip3] pytorch-triton==3.0.0+72734f086 [pip3] torch==2.6.0a0+df5bbc09d1.nv24.12 [pip3] torch_tensorrt==2.6.0a0 [pip3] torchprofile==0.0.4 [pip3] torchvision==0.20.0a0 [conda] Could not collect cc @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225 @Chillee @drisspg @yanboliang @BoyuanFeng
triaged,oncall: pt2,module: higher order operators,module: pt2-dispatcher,module: flex attention
low
Critical
2,805,914,111
godot
(4.4.beta1) Android app crashes when exported with .NET 9
### Tested versions **Reproducible:** 4.4.beta1 4.4.dev7 **Not Reproducible:** 4.4.dev3 ### System information Godot v4.4.beta1.mono - Windows 10 (build 19045) - OpenGL 3 (Compatibility) ### Issue description When exporting to Android with .NET 9 using Gradle Build via the Remote Deploy, the app crashes immediately at startup. Note: .NET 8 works fine, but .NET 9 causes the app to crash using the command at runtime: **adb logcat|grep godot** > 01-22 19:27:18.818 1895 3450 I ActivityTaskManager: START u0 {act=android.intent.action.MAIN flg=0x10000000 cmp=com.example.test_android/com.godot.game.GodotApp} from uid 2000 01-22 19:27:18.857 1895 2224 I ActivityManager: Start proc 1876:com.example.test_android/u0a411 for next-top-activity {com.example.test_android/com.godot.game.GodotApp} 01-22 19:27:19.885 1895 2006 W ActivityTaskManager: Force finishing activity com.example.test_android/com.godot.game.GodotApp 01-22 19:27:19.987 1895 2220 I GameServiceProviderInstance: Failed to remove task overlay. This is expected if the task is already destroyed: GameSessionRecord{mTaskId=306, mState=GAME_SESSION_ATTACHED, mRootComponentName=ComponentInfo{com.example.test_android/com.godot.game.GodotApp}, mIGameSession=android.service.games.IGameSession$Stub$Proxy@3a45eaa, mSurfacePackage=android.view.SurfaceControlViewHost$SurfacePackage@3504a9b} 01-22 19:27:20.386 1895 2200 W ActivityTaskManager: Activity top resumed state loss timeout for ActivityRecord{691de90 u0 com.example.test_android/com.godot.game.GodotApp} t-1 f}} ### Steps to reproduce 1. Godot 4.4.beta1 2. Create a new project 3. Create C# script and attach to a node in the main scene 4. Update the visual studio project to target .NET 9: ` <TargetFramework>net9.0</TargetFramework>` 4. Remote Deploy to Android with Gradle Build ### Minimal reproduction project (MRP) N/A
bug,platform:android,topic:dotnet,regression,topic:export
low
Critical
2,805,951,390
pytorch
Incomplete check of LR as a tensor in Optimizer
### 🐛 Describe the bug A tutorial "[Running the compiled optimizer with an LR Scheduler](https://pytorch.org/tutorials/recipes/compiling_optimizer_lr_scheduler.html)" presents how LR as a tensor is used. According to this tutorial, we should use a 0-dim tensor for LR. However, Optimizer can accept a 1-dim tensor of size 1. When using a 1-dim tensor, we get a runtime error. ``` optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda()) scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5) ``` ``` RuntimeError: fill_ only supports 0-dimension value tensor but got tensor with 1 dimensions. ``` Currently, Optimizer checks the tensor size: https://github.com/pytorch/pytorch/blob/faa10faa2cad1cf6eef95a3e5bda255be6bc4c87/torch/optim/adam.py#L50-L56 I think that Optimizer should check the tensor dimension: ``` if lr.dim() != 0: raise ValueError("Tensor lr must be 0-dimension") ``` The whole code I used: ``` import torch print(torch.__version__) # exit cleanly if we are on a device that doesn't support ``torch.compile`` if torch.cuda.get_device_capability() < (7, 0): print("Exiting because torch.compile is not supported on this device.") import sys sys.exit(0) # Setup random seed torch.manual_seed(12345) # Create simple model model = torch.nn.Sequential( *[torch.nn.Linear(1024, 1024, False, device="cuda") for _ in range(10)] ) input = torch.rand(1024, device="cuda") # run forward pass output = model(input) # run backward to populate the grads for our optimizer below output.sum().backward() optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda()) scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5) def fn(): optimizer.step() scheduler.step() opt_fn = torch.compile(fn) for i in range(5): opt_fn() print(i, optimizer.param_groups[0]['lr'].item()) ``` ### Error logs ``` --------------------------------------------------------------------------- RuntimeError Traceback (most recent call last) <ipython-input-2-67dacbf46b2f> in <cell line: 26>() 24 25 optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda()) ---> 26 scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5) 27 28 def fn(): /usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in __init__(self, optimizer, start_factor, end_factor, total_iters, last_epoch, verbose) 746 self.end_factor = end_factor 747 self.total_iters = total_iters --> 748 super().__init__(optimizer, last_epoch, verbose) 749 750 def get_lr(self): /usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in __init__(self, optimizer, last_epoch, verbose) 144 patch_track_step_called(self.optimizer) 145 self.verbose = _check_verbose_deprecated_warning(verbose) --> 146 self._initial_step() 147 148 def _initial_step(self): /usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in _initial_step(self) 149 """Initialize step counts and perform a step.""" 150 self._step_count = 0 --> 151 self.step() 152 153 def state_dict(self): /usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in step(self, epoch) 248 param_group, lr = data 249 if isinstance(param_group["lr"], Tensor): --> 250 param_group["lr"].fill_(lr) 251 else: 252 param_group["lr"] = lr RuntimeError: fill_ only supports 0-dimension value tensor but got tensor with 1 dimensions. ``` ### Versions PyTorch version: 2.5.1+cu121 Is debug build: False CUDA used to build PyTorch: 12.1 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.3 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: 14.0.0-1ubuntu1.1 CMake version: version 3.31.2 Libc version: glibc-2.35 Python version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-6.6.56+-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: 12.2.140 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: Tesla T4 GPU 1: Tesla T4 Nvidia driver version: 560.35.03 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.9.6 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.9.6 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 4 On-line CPU(s) list: 0-3 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) CPU @ 2.00GHz CPU family: 6 Model: 85 Thread(s) per core: 2 Core(s) per socket: 2 Socket(s): 1 Stepping: 3 BogoMIPS: 4000.30 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc cpuid tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch pti ssbd ibrs ibpb stibp fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves arat md_clear arch_capabilities Hypervisor vendor: KVM Virtualization type: full L1d cache: 64 KiB (2 instances) L1i cache: 64 KiB (2 instances) L2 cache: 2 MiB (2 instances) L3 cache: 38.5 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-3 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Mitigation; PTE Inversion Vulnerability Mds: Mitigation; Clear CPU buffers; SMT Host state unknown Vulnerability Meltdown: Mitigation; PTI Vulnerability Mmio stale data: Vulnerable: Clear CPU buffers attempted, no microcode; SMT Host state unknown Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Mitigation; IBRS Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP conditional; RSB filling; PBRSB-eIBRS Not affected; BHI SW loop, KVM SW loop Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Mitigation; Clear CPU buffers; SMT Host state unknown Versions of relevant libraries: [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.26.4 [pip3] nvidia-cublas-cu12==12.6.4.1 [pip3] nvidia-cuda-cupti-cu12==12.6.80 [pip3] nvidia-cuda-runtime-cu12==12.6.77 [pip3] nvidia-cudnn-cu12==9.6.0.74 [pip3] nvidia-cufft-cu12==11.3.0.4 [pip3] nvidia-curand-cu12==10.3.7.77 [pip3] nvidia-cusolver-cu12==11.7.1.2 [pip3] nvidia-cusparse-cu12==12.5.4.2 [pip3] nvidia-nccl-cu12==2.23.4 [pip3] nvidia-nvjitlink-cu12==12.6.85 [pip3] nvtx==0.2.10 [pip3] onnx==1.17.0 [pip3] optree==0.13.1 [pip3] pynvjitlink-cu12==0.4.0 [pip3] pytorch-ignite==0.5.1 [pip3] pytorch-lightning==2.5.0.post0 [pip3] torch==2.5.1+cu121 [pip3] torchaudio==2.5.1+cu121 [pip3] torchinfo==1.8.0 [pip3] torchmetrics==1.6.1 [pip3] torchsummary==1.5.1 [pip3] torchtune==0.5.0 [pip3] torchvision==0.20.1+cu121 [conda] Could not collect cc @vincentqb @jbschlosser @albanD @janeyx99 @crcrpar @chauhang @penguinwu
module: optimizer,triaged,actionable
low
Critical
2,805,959,485
vscode
Squash Merge will list a lot of commits msg that lead to crash in some cases
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.96.3 - OS Version: MacOS 14.5 (23F79) Steps to Reproduce: Assumes that have so many commits and need to merge into a branch with squash mode.
info-needed,git
low
Critical
2,805,981,004
flutter
Impeller Rendering failed
### Steps to reproduce The following problems occur when using texture createSurfaceProducer on Huawei p30 devices in flutter-3.27.3 environment: 1-Impeller surfaceProducer is not enabled?. setSize(width, width * 9 / 16) Setting texture size has no effect, Bitmap crashes using textured surface generation ![Image](https://github.com/user-attachments/assets/06b76e84-fc81-4eff-916b-d13056528a06) 2-Turn on Impeller render blur and second creation crashes ![Image](https://github.com/user-attachments/assets/5e3c82e8-869a-4d06-867b-02e5b5c36e9b) ### Expected results ... ### Actual results ... ### Code sample <details open><summary>Code sample</summary> ```dart private var width: Int = 0 if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.R) { val windowManager = this.flutterPluginBinding!!.applicationContext.getSystemService(Context.WINDOW_SERVICE) as WindowManager val windowMetrics = windowManager.currentWindowMetrics val bounds = windowMetrics.bounds width = bounds.width() } else { val metrics = DisplayMetrics() (this.flutterPluginBinding!!.applicationContext.getSystemService(Context.WINDOW_SERVICE) as WindowManager).defaultDisplay.getMetrics( metrics ) width = metrics.widthPixels } val createSurfaceProducer = this.flutterPluginBinding?.textureRegistry?.createSurfaceProducer() surfaceProducer = createSurfaceProducer surfaceProducer?.setSize(width, width * 9 / 16) ``` </details> ### Screenshots or Video The following problems occur when using texture createSurfaceProducer on Huawei p30 devices in flutter-3.27.3 environment: 1-Impeller surfaceProducer is not enabled?. setSize(width, width * 9 / 16) Setting texture size has no effect, Bitmap crashes using textured surface generation ![Image](https://github.com/user-attachments/assets/06b76e84-fc81-4eff-916b-d13056528a06) 2-Turn on Impeller render blur and second creation crashes ![Image](https://github.com/user-attachments/assets/5e3c82e8-869a-4d06-867b-02e5b5c36e9b) ### Logs <details open><summary>Logs</summary> ```console D/permissions_handler(19155): No android specific permissions needed for: 6 D/permissions_handler(19155): No permissions found in manifest for: []9 I/flutter (19155): [IMPORTANT:flutter/shell/platform/android/platform_view_android.cc(308)] Flutter recommends migrating plugins that create and register surface textures to the new surface producer API. See https://docs.flutter.dev/release/breaking-changes/android-surface-plugins V/version (19155): external version is 20240531_18:53:37_44aea9b4 I/AliFrameWork(19155): [0.9] [] :Ffmpeg version n4.3.1-8-gf381f263fe E/AliFrameWork(19155): [0.9] [JNI] :0328 JNI_OnLoad I/.saida.hengloo(19155): Thread[1,tid=19155,Native,Thread*=0x76e4c10800,peer=0x733ba540,"main"] recursive attempt to load library "/data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!/lib/arm64-v8a/libCicadaPlayer.so" I/AliFrameWork(19155): [0.9] [] :setWorkPath :/data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/lib/arm64/ I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0 I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0 W/Gralloc3(19155): allocator 3.x is not supported I/AliFrameWork(19155): [0.9] [] :AliyunCorePlayer jni_onVideoSizeChanged width = 2880 ,height = 1620 I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :mPtsDiscontinueDelta = 20000000 I/AliFrameWork(19155): [0.9] [AbrBufferAlgoStrategy] :BA already change to bitrate:0 E/AliFrameWork(19155): [0.9] [SMPAVDeviceManager] :config decoder error ret= -513 E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAudioDecoder error -514 E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAVPath SetUpAudioPath failed,url is rtmp://v-jiangyin3-rtmp-play.videiot.cn:1937/live/101532044-1?eyJrZXkiOjk5OTksInNpZ24iOiJPLWt3ZzU4WHV1UENrMS1NeVVsV2xsWm9YMkhqeVhBV2dHZERuWmhGWjg1UDg2RFpUcFhDTXFwUFNGV3NKZTB5Q1EwWFZBUTdtaWlQcUI3eHNhTEI1bUZOTFVtUWtqNk1FdUkyMTFqbko3SXh5dXhaaFZmUHBYYTNmdXdmc1hQUjljRXRYRDZkR2NXMVpFWUhDZGtJUlhMSjJLNWp1SmVVYkFVeUxoWTRrdTgzZ1BmQjVzVkRpX0RzM3NsRWNnUzdIMm1HdnZYd05NUUs2eWg1YzRQdVl4bThZRjNUT3NaQ2R3RVhBOFVwM2JzIn0 video decoder open error W/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :close audio stream E/AliFrameWork(19155): [0.9] [GLRender_OESContext] :compileShader mVertShader failed. ret = -1 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :function name eglPresentationTimeANDROID I/AliFrameWork(19155): [0.9] [GLRender egl_context] :sym 0x7767517bac, critical true I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL client major 1 minor 4 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL extension EGL_KHR_get_all_proc_addresses EGL_ANDROID_presentation_time EGL_KHR_swap_buffers_with_damage EGL_ANDROID_get_native_client_buffer EGL_ANDROID_front_buffer_auto_refresh EGL_ANDROID_get_frame_timestamps EGL_EXT_surface_SMPTE2086_metadata EGL_EXT_surface_CTA861_3_metadata EGL_KHR_image EGL_KHR_image_base EGL_EXT_image_gl_colorspace EGL_KHR_gl_colorspace EGL_KHR_gl_texture_2D_image EGL_KHR_gl_texture_cubemap_image EGL_KHR_gl_renderbuffer_image EGL_KHR_fence_sync EGL_KHR_create_context EGL_KHR_config_attribs EGL_KHR_surfaceless_context EGL_EXT_create_context_robustness EGL_ANDROID_image_native_buffer EGL_KHR_wait_sync EGL_ANDROID_recordable EGL_KHR_partial_update EGL_EXT_pixel_format_float EGL_KHR_mutable_render_buffer EGL_EXT_protected_content EGL_IMG_context_priority EGL_KHR_no_config_context I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL attr version 2 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL config num 1 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL choose best config I/AliFrameWork(19155): [0.9] [GLRender egl_context] :eglMakeCurrent: display(0x1) surface(0x764407ff80) context(0x764407ff00) I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext CreateSurface mDisplay 0x1 mConfig 0x76775a8008 window 0x764a5431f0 D/mali_winsys(19155): EGLint new_window_surface(egl_winsys_display *, void *, EGLSurface, EGLConfig, egl_winsys_surface **, EGLBoolean) returns 0x3000 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglCreateWindowSurface surface : 0x76327cf480 W/VideoCapabilities(19155): Unrecognized profile/level 0/0 for video/mpeg2 W/VideoCapabilities(19155): Unrecognized profile/level 0/2 for video/mpeg2 W/VideoCapabilities(19155): Unrecognized profile/level 0/3 for video/mpeg2 I/VideoCapabilities(19155): Unsupported profile 5 for video/mpeg2 I/chatty (19155): uid=10879(com.saida.henglook) ApsaraPlayerSer identical 2 lines I/VideoCapabilities(19155): Unsupported profile 5 for video/mpeg2 W/VideoCapabilities(19155): Unrecognized profile/level 1/32 for video/mp4v-es W/VideoCapabilities(19155): Unrecognized profile/level 32768/2 for video/mp4v-es W/VideoCapabilities(19155): Unrecognized profile/level 32768/64 for video/mp4v-es I/OMXClient(19155): IOmx service obtained I/ACodec (19155): In onAllocateComponent create compenent, codec name: OMX.hisi.video.decoder.avc D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason connectToSurface I/MediaCodec(19155): [OMX.hisi.video.decoder.avc] setting surface generation to 19614721 D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason connectToSurface(reconnect) D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason connectToSurface(reconnect) E/ACodec (19155): [OMX.hisi.video.decoder.avc] setPortMode on output to DynamicANWBuffer failed w/ err -2147483648 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): onStart I/HwExtendedUtils(19155): Set to window composer mode as 2 I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec) D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): set up nativeWindow 0x763bbba010 for 2880x1620, color 0x30d, rotation 0, usage 0x2900 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 6 buffers from a native window of size 7363584 on output port I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :send Frame mFlushState = 2. pts 2444137047000 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Now handling output port settings change I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now disabled. I/HwExtendedUtils(19155): Set to window composer mode as 2 I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec) D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): set up nativeWindow 0x763bbba010 for 2880x1632, color 0x30d, rotation 0, usage 0x2900 W/ACodec (19155): [OMX.hisi.video.decoder.avc] setting nBufferCountActual to 11 failed: -1010 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 10 buffers from a native window of size 7050240 on output port I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now reenabled. D/ACodec (19155): sendVideoFpsDataToiAware time:2444137047000 fps:-1 msg:-1 I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player NotifyFirstFrame D/ACodec (19155): sendVideoFpsDataToiAware time:2444137087000 fps:25 msg:25 I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA V/AudioManager(19155): querySoundEffectsEnabled... E/flutter (19155): [ERROR:flutter/shell/platform/android/surface_texture_external_texture_vk_impeller.cc(122)] Break on 'ImpellerValidationBreak' to inspect point of failure: Invalid external texture. I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player ReadPacket Stop I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :not opened I/AliFrameWork(19155): [0.9] [MeidaPlayerUtil] :KPI test finish: total fps:23.1 I/AliFrameWork(19155): [0.9] [GLRender] :drop a frame pts = 2444140407000 E/AliFrameWork(19155): [0.9] [GLRender] :~GLRender W/BufferQueueProducer(19155): [SurfaceTexture-0-19155-0]:1390: disconnect: not connected (req=1) W/libEGL (19155): EGLNativeWindowType 0x764568c010 disconnect failed I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglDestroySurface eglSurface 0x76327cf480 I/MediaCodec(19155): kWhatFlush processing is complete I/ACodec (19155): [OMX.hisi.video.decoder.avc] ExecutingState flushing now (codec owns 2/5 input, 5/10 output). I/MediaCodec(19155): Start Processing kWhatFlushCompleted I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :clearCache. ret 0, flush state 1 W/ACodec (19155): forcing OMX state to Idle when received shutdown in ExecutingState I/MediaCodec(19155): start process kWhatStopCompleted D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason disconnectFromSurface ======== Exception caught by image resource service ================================================ The following HttpExceptionWithStatus was thrown resolving an image codec: HttpException: Invalid statusCode: 400, uri = https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0 When the exception was thrown, this was the stack: Image provider: CachedNetworkImageProvider("https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0", scale: 1.0) Image key: CachedNetworkImageProvider("https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0", scale: 1.0) ==================================================================================================== D/permissions_handler(19155): No android specific permissions needed for: 6 D/permissions_handler(19155): No permissions found in manifest for: []9 I/flutter (19155): [IMPORTANT:flutter/shell/platform/android/platform_view_android.cc(308)] Flutter recommends migrating plugins that create and register surface textures to the new surface producer API. See https://docs.flutter.dev/release/breaking-changes/android-surface-plugins I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0 I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0 I/AliFrameWork(19155): [0.9] [] :AliyunCorePlayer jni_onVideoSizeChanged width = 2880 ,height = 1620 I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :mPtsDiscontinueDelta = 20000000 E/AliFrameWork(19155): [0.9] [SMPAVDeviceManager] :config decoder error ret= -513 E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAudioDecoder error -514 E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAVPath SetUpAudioPath failed,url is rtmp://v-jiangyin3-rtmp-play.videiot.cn:1937/live/101532044-1?eyJrZXkiOjk5OTksInNpZ24iOiJPLWt3ZzU4WHV1UENrMS1NeVVsV2xsWm9YMkhqeVhBV2dHZERuWmhGWjg0SEtqSkFzbnBrWXNySUQxOThYcm5PcENTMnlxbllBd1NKcGN4c3liNjFzMmU3ZWFHNFl6SkhaZXB3aXl2S18tYk51RkZUajJzRWMzRGgtRkVGX3Z5RjRlQ2E2NlJvZFBqaVE5UEozcWdMM1p4LTc3bnYzTEJ0MzdqX1NDVDhfejAxMTJ3YWMxUFdaY3Zhc1p5R3pVdWE3MWJSMHNJSzRYbWhpVzREN3ptaUE5QU1xVGZnM3VJMHU2dW5uQTBHQVpRIn0 video decoder open error W/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :close audio stream I/AliFrameWork(19155): [0.9] [AbrBufferAlgoStrategy] :BA already change to bitrate:0 E/AliFrameWork(19155): [0.9] [GLRender_OESContext] :compileShader mVertShader failed. ret = -1 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :Egl .so already loaded I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL client major 1 minor 4 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL extension EGL_KHR_get_all_proc_addresses EGL_ANDROID_presentation_time EGL_KHR_swap_buffers_with_damage EGL_ANDROID_get_native_client_buffer EGL_ANDROID_front_buffer_auto_refresh EGL_ANDROID_get_frame_timestamps EGL_EXT_surface_SMPTE2086_metadata EGL_EXT_surface_CTA861_3_metadata EGL_KHR_image EGL_KHR_image_base EGL_EXT_image_gl_colorspace EGL_KHR_gl_colorspace EGL_KHR_gl_texture_2D_image EGL_KHR_gl_texture_cubemap_image EGL_KHR_gl_renderbuffer_image EGL_KHR_fence_sync EGL_KHR_create_context EGL_KHR_config_attribs EGL_KHR_surfaceless_context EGL_EXT_create_context_robustness EGL_ANDROID_image_native_buffer EGL_KHR_wait_sync EGL_ANDROID_recordable EGL_KHR_partial_update EGL_EXT_pixel_format_float EGL_KHR_mutable_render_buffer EGL_EXT_protected_content EGL_IMG_context_priority EGL_KHR_no_config_context I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL attr version 2 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL config num 1 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL choose best config I/AliFrameWork(19155): [0.9] [GLRender egl_context] :eglMakeCurrent: display(0x1) surface(0x76327cfd80) context(0x76327cf900) I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext CreateSurface mDisplay 0x1 mConfig 0x76775a8008 window 0x764571ec28 D/mali_winsys(19155): EGLint new_window_surface(egl_winsys_display *, void *, EGLSurface, EGLConfig, egl_winsys_surface **, EGLBoolean) returns 0x3000 I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglCreateWindowSurface surface : 0x76327cff00 I/OMXClient(19155): IOmx service obtained I/ACodec (19155): In onAllocateComponent create compenent, codec name: OMX.hisi.video.decoder.avc D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason connectToSurface I/MediaCodec(19155): [OMX.hisi.video.decoder.avc] setting surface generation to 19614722 D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason connectToSurface(reconnect) D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason connectToSurface(reconnect) E/ACodec (19155): [OMX.hisi.video.decoder.avc] setPortMode on output to DynamicANWBuffer failed w/ err -2147483648 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): onStart I/HwExtendedUtils(19155): Set to window composer mode as 2 I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec) D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): set up nativeWindow 0x763a818010 for 2880x1620, color 0x30d, rotation 0, usage 0x2900 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 6 buffers from a native window of size 7363584 on output port I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :send Frame mFlushState = 2. pts 2444150007000 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR) I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Now handling output port settings change I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now disabled. I/HwExtendedUtils(19155): Set to window composer mode as 2 I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec) D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage D/SurfaceUtils(19155): set up nativeWindow 0x763a818010 for 2880x1632, color 0x30d, rotation 0, usage 0x2900 W/ACodec (19155): [OMX.hisi.video.decoder.avc] setting nBufferCountActual to 11 failed: -1010 I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 10 buffers from a native window of size 7050240 on output port I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now reenabled. D/ACodec (19155): sendVideoFpsDataToiAware time:2444150007000 fps:60 msg:60 I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player NotifyFirstFrame I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [GLRender] :drop a frame pts = 2444161687000 I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA W/System (19155): A resource failed to call end. I/chatty (19155): uid=10879(com.saida.henglook) FinalizerDaemon identical 1 line W/System (19155): A resource failed to call end. F/libc (19155): Fatal signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0x660 in tid 19172 (FinalizerDaemon), pid 19155 (.saida.henglook) *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** Build fingerprint: 'HUAWEI/ELE-AL00/HWELE:10/HUAWEIELE-AL00/10.1.0.162C00:user/release-keys' Revision: '0' ABI: 'arm64' SYSVMTYPE: Maple APPVMTYPE: Art Timestamp: 2025-01-23 13:42:49+0800 pid: 19155, tid: 19172, name: FinalizerDaemon >>> com.saida.henglook <<< uid: 10879 signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0x660 Cause: null pointer dereference x0 0000000000000660 x1 00000076d8dbb5d4 x2 000000006ff4fb74 x3 00000076cd45b000 x4 00000076d8dbc660 x5 00000076e3b12ddf x6 647260646b647116 x7 7f7f7f7f7f7f7f7f x8 5fae25f0c64e4ce6 x9 5fae25f0c64e4ce6 x10 0000000000430000 x11 00000076e4bfc000 x12 00000076e473e224 x13 00000076e473e26c x14 00000076e473e2cc x15 0000000000000000 x16 000000776863dab0 x17 0000007769029250 x18 000000767e8c8000 x19 0000000000000660 x20 0000000000000000 x21 00000076cd45b000 x22 00000076d8dbb810 x23 00000076e3b12ddf x24 0000000000000004 x25 00000076d8dbd020 x26 00000076cd45b0b0 x27 0000000000000001 x28 0000000000000000 x29 00000076d8dbb580 sp 00000076d8dbb570 lr 00000077685edfc4 pc 0000007769029250 backtrace: #00 pc 00000000000d0250 /apex/com.android.runtime/lib64/bionic/libc.so (pthread_mutex_lock) (BuildId: b91c775ccc9b0556e91bc575a2511cd0) #01 pc 00000000000acfc0 /system/lib64/libgui.so (android::ConsumerBase::abandon()+24) (BuildId: 38bb0c19e01050d5b9cf5882b40316d7) #02 pc 00000000001c4b30 /system/lib64/libandroid_runtime.so (android::SurfaceTexture_release(_JNIEnv*, _jobject*)+80) (BuildId: a92c351fcfda20f10ff25c04ed924f5f) #03 pc 00000000002e649c /system/framework/arm64/boot-framework.oat (art_jni_trampoline+124) (BuildId: 6e4646237361a21448639f9cbe5b64c978005ca6) #04 pc 0000000000147334 /apex/com.android.runtime/lib64/libart.so (art_quick_invoke_stub+548) (BuildId: 617b4481d2d525502ced0629cac33c57) #05 pc 00000000001561b4 /apex/com.android.runtime/lib64/libart.so (art::ArtMethod::Invoke(art::Thread*, unsigned int*, unsigned int, art::JValue*, char const*)+252) (BuildId: 617b4481d2d525502ced0629cac33c57) #06 pc 00000000002fd900 /apex/com.android.runtime/lib64/libart.so (art::interpreter::ArtInterpreterToCompiledCodeBridge(art::Thread*, art::ArtMethod*, art::ShadowFrame*, unsigned short, art::JValue*)+384) (BuildId: 617b4481d2d525502ced0629cac33c57) #07 pc 00000000002f8bd0 /apex/com.android.runtime/lib64/libart.so (bool art::interpreter::DoCall<false, false>(art::ArtMethod*, art::Thread*, art::ShadowFrame&, art::Instruction const*, unsigned short, art::JValue*)+912) (BuildId: 617b4481d2d525502ced0629cac33c57) #08 pc 00000000005cd8fc /apex/com.android.runtime/lib64/libart.so (MterpInvokeDirect+400) (BuildId: 617b4481d2d525502ced0629cac33c57) #09 pc 0000000000141914 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_direct+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #10 pc 00000000003cc2b4 /system/framework/framework.jar (android.graphics.SurfaceTexture.release) #11 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57) #12 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #13 pc 00000000004eccba [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureWrapper.release+14) #14 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57) #15 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #16 pc 00000000004ebc7e [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.FlutterRenderer$SurfaceTextureRegistryEntry.release+78) #17 pc 00000000005cd060 /apex/com.android.runtime/lib64/libart.so (MterpInvokeInterface+1752) (BuildId: 617b4481d2d525502ced0629cac33c57) #18 pc 0000000000141a14 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_interface+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #19 pc 00000000004ecab8 [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureSurfaceProducer.release+4) #20 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57) #21 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #22 pc 00000000004eca58 [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureSurfaceProducer.finalize+16) #23 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57) #24 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #25 pc 00000000001b45aa /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$FinalizerDaemon.doFinalize+22) #26 pc 00000000005cdbfc /apex/com.android.runtime/lib64/libart.so (MterpInvokeDirect+1168) (BuildId: 617b4481d2d525502ced0629cac33c57) #27 pc 0000000000141914 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_direct+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #28 pc 00000000001b469c /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$FinalizerDaemon.runInternal+164) #29 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57) #30 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #31 pc 00000000001b439e /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$Daemon.run+50) #32 pc 00000000005cd060 /apex/com.android.runtime/lib64/libart.so (MterpInvokeInterface+1752) (BuildId: 617b4481d2d525502ced0629cac33c57) #33 pc 0000000000141a14 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_interface+20) (BuildId: 617b4481d2d525502ced0629cac33c57) #34 pc 00000000000eabdc /apex/com.android.runtime/javalib/core-oj.jar (java.lang.Thread.run+8) #35 pc 00000000002ce22c /apex/com.android.runtime/lib64/libart.so (_ZN3art11interpreterL7ExecuteEPNS_6ThreadERKNS_20CodeItemDataAccessorERNS_11ShadowFrameENS_6JValueEbb.llvm.10887373532384510885+320) (BuildId: 617b4481d2d525502ced0629cac33c57) #36 pc 00000000005bc090 /apex/com.android.runtime/lib64/libart.so (artQuickToInterpreterBridge+1012) (BuildId: 617b4481d2d525502ced0629cac33c57) #37 pc 0000000000150468 /apex/com.android.runtime/lib64/libart.so (art_quick_to_interpreter_bridge+88) (BuildId: 617b4481d2d525502ced0629cac33c57) #38 pc 0000000000147334 /apex/com.android.runtime/lib64/libart.so (art_quick_invoke_stub+548) (BuildId: 617b4481d2d525502ced0629cac33c57) #39 pc 00000000001561b4 /apex/com.android.runtime/lib64/libart.so (art::ArtMethod::Invoke(art::Thread*, unsigned int*, unsigned int, art::JValue*, char const*)+252) (BuildId: 617b4481d2d525502ced0629cac33c57) #40 pc 00000000004d8820 /apex/com.android.runtime/lib64/libart.so (art::(anonymous namespace)::InvokeWithArgArray(art::ScopedObjectAccessAlreadyRunnable const&, art::ArtMethod*, art::(anonymous namespace)::ArgArray*, art::JValue*, char const*)+104) (BuildId: 617b4481d2d525502ced0629cac33c57) #41 pc 00000000004d98b4 /apex/com.android.runtime/lib64/libart.so (art::InvokeVirtualOrInterfaceWithJValues(art::ScopedObjectAccessAlreadyRunnable const&, _jobject*, _jmethodID*, jvalue const*)+416) (BuildId: 617b4481d2d525502ced0629cac33c57) #42 pc 000000000051ca8c /apex/com.android.runtime/lib64/libart.so (art::Thread::CreateCallback(void*)+1232) (BuildId: 617b4481d2d525502ced0629cac33c57) #43 pc 00000000000cf7c0 /apex/com.android.runtime/lib64/bionic/libc.so (__pthread_start(void*)+36) (BuildId: b91c775ccc9b0556e91bc575a2511cd0) #44 pc 00000000000721a8 /apex/com.android.runtime/lib64/bionic/libc.so (__start_thread+64) (BuildId: b91c775ccc9b0556e91bc575a2511cd0) Lost connection to device. ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [!] Flutter (Channel stable, 3.27.3, on Microsoft Windows [版本 10.0.22631.4751], locale zh-CN) • Flutter version 3.27.3 on channel stable at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter ! Warning: `flutter` on your path resolves to D:\Android\AndroidSdk\flutter_windows_3.24.3-stable\flutter\bin\flutter, which is not inside your current Flutter SDK checkout at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter. Consider adding D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter\bin to the front of your path. ! Warning: `dart` on your path resolves to D:\Android\AndroidSdk\flutter_windows_3.24.3-stable\flutter\bin\dart, which is not inside your current Flutter SDK checkout at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter. Consider adding D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter\bin to the front of your path. • Upstream repository https://github.com/flutter/flutter.git • Framework revision c519ee916e (35 hours ago), 2025-01-21 10:32:23 -0800 • Engine revision e672b006cb • Dart version 3.6.1 • DevTools version 2.40.2 • If those were intentional, you can disregard the above warnings; however it is recommended to use "git" directly to perform update checks and upgrades. [√] Windows Version (Installed version of Windows is version 10 or higher) [!] Android toolchain - develop for Android devices (Android SDK version 34.0.0) • Android SDK at D:\Android\AndroidSdk X cmdline-tools component is missing Run `path/to/sdkmanager --install "cmdline-tools;latest"` See https://developer.android.com/studio/command-line for more details. X Android license status unknown. Run `flutter doctor --android-licenses` to accept the SDK licenses. See https://flutter.dev/to/windows-android-setup for more details. [√] Chrome - develop for the web • Chrome at C:\Users\wjia3\AppData\Local\Google\Chrome\Application\chrome.exe [√] Visual Studio - develop Windows apps (Visual Studio 生成工具 2019 16.11.5) • Visual Studio at C:\Program Files (x86)\Microsoft Visual Studio\2019\BuildTools • Visual Studio 生成工具 2019 version 16.11.31729.503 • Windows 10 SDK version 10.0.19041.0 [√] Android Studio (version 2024.1) • Android Studio at D:\Program Files\Android\Android Studio • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.10+0--11609105) [√] Android Studio (version 2024.2) • Android Studio at D:\Program Files\Android\Android Studio1 • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.10+0--11609105) [√] IntelliJ IDEA Ultimate Edition (version 2022.1) • IntelliJ at D:\Program Files\JetBrains\IntelliJ IDEA 2022.1.4 • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart [√] Connected device (4 available) • ELE AL00 (mobile) • Q5S5T19603021160 • android-arm64 • Android 10 (API 29) • Windows (desktop) • windows • windows-x64 • Microsoft Windows [版本 10.0.22631.4751] • Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83 • Edge (web) • edge • web-javascript • Microsoft Edge 132.0.2957.115 [!] Network resources X A network error occurred while checking "https://maven.google.com/": 信号灯超时时间已到 X A network error occurred while checking "https://github.com/": 信号灯超时时间已到 ! Doctor found issues in 3 categories. Process finished with exit code 0 ``` </details>
waiting for customer response,in triage
low
Critical
2,805,989,403
go
x/tools/go/ssa: TestGenericFunctionSelector fails with EADDR (netbsd)
``` #!watchflakes default <- pkg == "golang.org/x/tools/go/ssa" && test == "TestGenericFunctionSelector" ``` Issue created automatically to collect these failures. Example ([log](https://ci.chromium.org/b/8725122664788146929)): === RUN TestGenericFunctionSelector builder_test.go:799: err: fork/exec /home/swarming/.swarming/w/ir/x/w/goroot/bin/go: bad address: stderr: --- FAIL: TestGenericFunctionSelector (0.74s) — [watchflakes](https://go.dev/wiki/Watchflakes)
NeedsInvestigation,Tools
low
Critical
2,805,997,681
go
cmd/compile: worse performance when built with pgo
### Go version go version go1.23.5 linux/amd64 ### Output of `go env` in your module/workspace: ```shell GO111MODULE='' GOARCH='amd64' GOBIN='' GOCACHE='/home/transaction/.cache/go-build' GOENV='/home/transaction/.config/go/env' GOEXE='' GOEXPERIMENT='' GOFLAGS='' GOHOSTARCH='amd64' GOHOSTOS='linux' GOINSECURE='' GOMODCACHE='/home/transaction/go/pkg/mod' GONOPROXY='' GONOSUMDB='' GOOS='linux' GOPATH='/home/transaction/go' GOPRIVATE='' GOPROXY='https://proxy.golang.org,direct' GOROOT='/home/transaction/go/pkg/mod/golang.org/[email protected]' GOSUMDB='sum.golang.org' GOTMPDIR='' GOTOOLCHAIN='auto' GOTOOLDIR='/home/transaction/go/pkg/mod/golang.org/[email protected]/pkg/tool/linux_amd64' GOVCS='' GOVERSION='go1.23.5' GODEBUG='' GOTELEMETRY='local' GOTELEMETRYDIR='/home/transaction/.config/go/telemetry' GCCGO='gccgo' GOAMD64='v1' AR='ar' CC='gcc' CXX='g++' CGO_ENABLED='1' GOMOD='/home/transaction/tidb/go.mod' GOWORK='' CGO_CFLAGS='-O2 -g' CGO_CPPFLAGS='' CGO_CXXFLAGS='-O2 -g' CGO_FFLAGS='-O2 -g' CGO_LDFLAGS='-O2 -g' PKG_CONFIG='pkg-config' GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -ffile-prefix-map=/tmp/go-build338251587=/tmp/go-build -gno-record-gcc-switches' ``` ### What did you do? Building with [PGO](https://go.dev/doc/pgo), but result in performance regressions. The build flag is: ``` CGO_ENABLED=1 GO111MODULE=on go build -tags codes -pgo=default.pgo -ldflags '-X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v9.0.0-alpha-206-gaccc969cd5" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBBuildTS=2025-01-23 06:01:16" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBGitHash=accc969cd5e95a992368249fcf86f11887e7b11a" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBGitBranch=pgo-read-only-dev5" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBEdition=Community" ' -o bin/tidb-server ./cmd/tidb-server ``` Then start the TiDB and run sysbench test, following is the test result, the QPS of building with PGO is lower. workload | thread | Version | QPS -- | -- | -- | -- batch_get | 64 | without-pgo | 23165 batch_get | 64 | with-pgo | 21077 </byte-sheet-html-origin><!--EndFragment--> </byte-sheet-html-origin><!--EndFragment--> ### What did you see happen? The performance of build with PGO is worse because `newstack` appeared during execution. [profile.zip](https://github.com/user-attachments/files/18515784/profile.zip) CPU profile of building without PGO: <img width="3006" alt="Image" src="https://github.com/user-attachments/assets/70a9f8f5-921f-4264-8ca6-1d777d9b20ae" /> CPU profile of building with PGO: <img width="3003" alt="Image" src="https://github.com/user-attachments/assets/d401353e-db38-4c6e-ba2b-9cae189ddf55" /> ### What did you expect to see? no `newstack` during execution. Is there any way to avoid `newstack` during execution, such as set stack size when creating goroutine?
NeedsInvestigation,compiler/runtime,BugReport
low
Critical
2,806,003,966
PowerToys
在edge上无法反馈
### Provide a description of requested docs changes 在标题栏输入任意键会跳转至出现错误页面
Issue-Docs,Needs-Triage
low
Minor
2,806,004,012
PowerToys
在edge上无法反馈
### Provide a description of requested docs changes 在标题栏输入任意键会跳转至出现错误页面
Issue-Docs,Needs-Triage
low
Minor
2,806,008,166
PowerToys
鼠标荧光笔无法置于 always on top上
### Provide a description of requested docs changes 在置顶窗口后鼠标移至置顶的窗口会导致荧光笔无法使用,而在置顶窗口之外的区域依旧能够使用
Issue-Docs,Needs-Triage
low
Minor
2,806,059,279
pytorch
Mark Dynamic does not work for nn module constructor inputs
for the following code if we compile it we get the following graph : code: ``` import torch import torch._dynamo.config from torch.utils.checkpoint import _is_compiling @torch.compile() class Y(torch.nn.Module): def __init__(self): super().__init__() def forward(self, x): return x.view(-1, self._N_input, 192) @torch.compile(dynamic=True) class M(torch.nn.Module): def __init__(self, input): super().__init__() self._N_input = input.size()[0] def forward(self, x, z): # input is [B, n * d] x = x*2 x = x*5 x = x *3 y = Y() y._N_input = self._N_input x = y.forward(x)# [B, n, d] x = x*20 x = x*30 x = x*43 return x x = torch.randn(5, 3210, 192) # Shape [B=4, n*d=12] num_inputs = torch.randn(3210) m = M(num_inputs) y1, z1 = 3210, 192 # y1 * z1 == 12 output1 = m(x, z1) print(output1) ``` graph: Note L_self_N_input is Sym(s1) ``` V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] TRACED GRAPH V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ===== __compiled_fn_1 ===== V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] /home/lsakka/pytorch/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] def forward(self, s0: "Sym(s0)", s1: "Sym(s1)", L_x_: "f32[s0, s1, 192][192*s1, 192, 1]cpu", L_self_N_input: "Sym(s1)"): V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_x_ = L_x_ V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_self_n_input = L_self_N_input V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:751 in forward, code: x = x*2 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = l_x_ * 2; l_x_ = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:752 in forward, code: x = x*5 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_1: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x * 5; x = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:753 in forward, code: x = x *3 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_2: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_1 * 3; x_1 = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:737 in __init__, code: super().__init__() V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] _log_api_usage_once = torch._C._log_api_usage_once('python.nn_module'); _log_api_usage_once = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:740 in forward, code: return x.view(-1, self._N_input, 192) V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_3: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_2.view(-1, l_self_n_input, 192); x_2 = l_self_n_input = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:759 in forward, code: x = x*20 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_4: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_3 * 20; x_3 = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:760 in forward, code: x = x*30 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_5: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_4 * 30; x_4 = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:761 in forward, code: x = x*43 V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_6: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_5 * 43; x_5 = None V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] return (x_6,) V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ``` Now if we want to use mark_dynamic instead, I : 1) remove(dynamic=True) 2) call mark dynamic as the following: ``` mport torch import torch._dynamo.config from torch.utils.checkpoint import _is_compiling @torch.compile() class Y(torch.nn.Module): def __init__(self): super().__init__() def forward(self, x): return x.view(-1, self._N_input, 192) @torch.compile() class M(torch.nn.Module): def __init__(self, input): super().__init__() self._N_input = input.size()[0] def forward(self, x, z): # input is [B, n * d] x = x*2 x = x*5 x = x *3 y = Y() y._N_input = self._N_input x = y.forward(x)# [B, n, d] x = x*20 x = x*30 x = x*43 return x x = torch.randn(5, 3210, 192) # Shape [B=4, n*d=12] num_inputs = torch.randn(3210) torch._dynamo.decorators.mark_dynamic(x, 1) torch._dynamo.decorators.mark_dynamic(num_inputs, 0) m = M(num_inputs) y1, z1 = 3210, 192 # y1 * z1 == 12 output1 = m(x, z1) print(output1) ``` but this generate this graph, looks like mark_dynamic did not work on num_inputs raise ConstraintViolationError( torch.fx.experimental.symbolic_shapes.ConstraintViolationError: Constraints violated (L['x'].size()[1])! For more information, run with TORCH_LOGS="+dynamic". - Not all values of RelaxedUnspecConstraint(L['x'].size()[1]) are valid because L['x'].size()[1] was inferred to be a constant (3210). ``` V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] TRACED GRAPH V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ===== __compiled_fn_1 ===== V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] /home/lsakka/pytorch/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] def forward(self, L_x_: "f32[5, 3210, 192][616320, 192, 1]cpu"): V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_x_ = L_x_ V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:751 in forward, code: x = x*2 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x: "f32[5, 3210, 192][616320, 192, 1]cpu" = l_x_ * 2; l_x_ = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:752 in forward, code: x = x*5 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_1: "f32[5, 3210, 192][616320, 192, 1]cpu" = x * 5; x = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:753 in forward, code: x = x *3 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_2: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_1 * 3; x_1 = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:737 in __init__, code: super().__init__() V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] _log_api_usage_once = torch._C._log_api_usage_once('python.nn_module'); _log_api_usage_once = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:740 in forward, code: return x.view(-1, self._N_input, 192) V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_3: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_2.view(-1, 3210, 192); x_2 = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:759 in forward, code: x = x*20 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_4: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_3 * 20; x_3 = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:760 in forward, code: x = x*30 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_5: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_4 * 30; x_4 = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:761 in forward, code: x = x*43 V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_6: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_5 * 43; x_5 = None V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] return (x_6,) V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ``` cc @chauhang @penguinwu @ezyang @bobrenjc93
triaged,module: dynamic shapes
low
Critical
2,806,171,351
react
Title Unnecessary updates to defaultValue and checked properties in <textarea> and <input> elements
### Description It looks like React is updating the defaultValue of <textarea> elements and the checked property of <input> elements even when their values haven’t changed. This leads to unnecessary DOM mutations, which can impact performance, especially when there are many components that re-render frequently. ### Steps to Reproduce Render a controlled <textarea> or <input> with defaultValue or checked set. Trigger a re-render by updating some other unrelated props. Notice that React still updates the defaultValue or checked properties, even though the values haven't changed. ### Expected Behavior React should only update the defaultValue or checked properties when the values actually change, avoiding unnecessary DOM updates. ### Actual Behavior The properties are being updated unnecessarily, even if the values remain unchanged, leading to redundant DOM mutations. ### Suggested Solution React should add a check to only update the defaultValue and checked properties if their values have changed. This would reduce unnecessary updates and improve overall performance. ### Environment React version: [Insert React version] Browser: [Insert browser and version] Operating System: [Insert OS name and version] ### Key Words / Search Terms defaultValue, checked, textarea, input, performance, DOM mutations, unnecessary updates ### Repository Version [v19.0.0](https://github.com/facebook/react/releases/tag/v19.0.0)
Status: Unconfirmed
medium
Major
2,806,172,873
react
Bug:
The custom Jest matchers introduced for async operations in the React Test Renderer (toFlushAll, toFlushThrough, toFlushAndThrow, toClearYields) throw unexpected errors in specific scenarios involving yielded values. These errors prevent the test suite from running as expected when handling complex async updates. React version: 18.2.0 ## Steps To Reproduce 1.Install React Test Renderer and add the custom Jest matchers. 2.Write a test using the toFlushAll matcher on a component rendering async updates. 3.Run the test suite and observe the unexpected behavior during flushing. Link to code example: link : https://codesandbox.io/p/sandbox/new code: import React from "react"; import TestRenderer, { act } from "react-test-renderer"; describe("React Test Renderer Async Matchers", () => { it("throws an unexpected error with toFlushAll", () => { const MyComponent = () => { const [state, setState] = React.useState(0); React.useEffect(() => { setTimeout(() => setState(1), 1000); }, []); return <div>{state}</div>; }; const renderer = TestRenderer.create(<MyComponent />); expect(() => { act(() => { jest.runAllTimers(); renderer.unstable_flushAll(); // Fails here with an unexpected error }); }).not.toThrow(); }); }); ## The current behavior When using toFlushAll, the test fails unexpectedly with a TypeError, preventing it from completing the async operations. The issue occurs when the matcher processes components with multiple nested yields. ## The expected behavior The toFlushAll matcher should flush all pending async updates and yield the correct values without throwing errors. It should handle components with nested async updates seamlessly.
Status: Unconfirmed
medium
Critical
2,806,200,871
flutter
[native assets] [testing] Cleanup fake in tests
I stumbled upon some weird looking code in some of our testing fakes: https://github.com/flutter/flutter/blob/97ca57cf0823a1948cbb6646a04054550842080d/packages/flutter_tools/test/general.shard/isolated/fake_native_assets_build_runner.dart#L25-L26 https://github.com/flutter/flutter/blob/97ca57cf0823a1948cbb6646a04054550842080d/packages/flutter_tools/test/general.shard/isolated/fake_native_assets_build_runner.dart#L37-L68 `onBuild` takes a `BuildConfig` (single hook invocation), but outputs a `BuildResult` which is the aggregated of all hook invocations. The implementation then does a loop over all packages and simply overwrites the `result` if there is more than one package with native assets. Unsurprisingly, we don't have any invocations with more than one package with native assets. This should be rewritten in one of the two following ways: * `final BuildOutput? Function(BuildConfig)? onBuild;` and loop over all packages and combine the results, or * `final BuildResult? Function()? onBuild` and don't loop over the packages. If this should be a shallow mock, it should just record what it is invoked with and not do anything fancy. So I'm leaning towards option 2. cc @mkustermann you added the loop over the packages (presumably to get access to a `BuildConfig`), any opinions?
P3,c: tech-debt,team-tool,triaged-tool
low
Minor
2,806,204,015
kubernetes
code-generator v1.32.x defaulter-gen error
### What happened? defaulter-gen `zz_generated.defaults.go` Content error. example case: https://github.com/dongjiang1989/customapis/tree/main/pkg/apis/custom/v1 ```yaml package v1 import ( corev1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" ) // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // TestType is a top-level type. A client is created for it. type TestType struct { metav1.TypeMeta `json:",inline"` // +optional metav1.ObjectMeta `json:"metadata,omitempty"` // +optional Spec TestSpec `json:"spec,omitempty"` } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // TestTypeList is a top-level list type. The client methods for lists are automatically created. // You are not supposed to create a separated client for this one. type TestTypeList struct { metav1.TypeMeta `json:",inline"` // +optional metav1.ListMeta `json:"metadata,omitempty"` Items []TestType `json:"items"` } type TestSpec struct { ReplicaSpecs map[string]*ReplicaSpec `json:"replicaSpecs"` --------> this part } // +k8s:openapi-gen=true // +k8s:deepcopy-gen=true type ReplicaSpec struct { // +optional Replicas *int32 `json:"replicas,omitempty"` Template corev1.PodTemplateSpec `json:"template,omitempty"` } ```` output: ```yaml package v1 import ( corev1 "k8s.io/api/core/v1" ---------> This imported package is not required runtime "k8s.io/apimachinery/pkg/runtime" ) // RegisterDefaults adds defaulters functions to the given scheme. // Public to allow building arbitrary schemes. // All generated defaulters are covering - they call all nested defaulters. func RegisterDefaults(scheme *runtime.Scheme) error { scheme.AddTypeDefaultingFunc(&TestType{}, func(obj interface{}) { SetObjectDefaults_TestType(obj.(*TestType)) }) scheme.AddTypeDefaultingFunc(&TestTypeList{}, func(obj interface{}) { SetObjectDefaults_TestTypeList(obj.(*TestTypeList)) }) return nil } func SetObjectDefaults_TestType(in *TestType) { SetDefaults_TestType(in) } func SetObjectDefaults_TestTypeList(in *TestTypeList) { for i := range in.Items { a := &in.Items[i] SetObjectDefaults_TestType(a) } } ``` ### What did you expect to happen? defaulter-gen `zz_generated.defaults.go` Content error. ### How can we reproduce it (as minimally and precisely as possible)? ```bash dongjiang@MacBook Pro:customapis $ pwd /Users/dongjiang/Documents/go/src/github.com/dongjiang1989/customapis dongjiang@MacBook Pro:customapis $ ./hack/update-codegen.sh /Users/dongjiang/Documents/go/pkg/mod/k8s.io/[email protected] Generating deepcopy code for 1 targets Generating defaulter code for 1 targets Generating openapi code for 1 targets Generating applyconfig code for 1 targets Generating client code for 1 targets Generating lister code for 1 targets Generating informer code for 1 targets ``` ### Anything else we need to know? _No response_ ### Kubernetes version <details> ```console $ kubectl version # paste output here ``` </details> code-generator v1.32.1 and v1.32.0 ### Cloud provider <details> </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release # paste output here $ uname -a # paste output here # On Windows: C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture # paste output here ``` </details> ### Install tools <details> </details> ### Container runtime (CRI) and version (if applicable) <details> </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details>
kind/bug,needs-sig,needs-triage
low
Critical
2,806,213,306
opencv
“cv2.CAP_PROP_POS_FRAMES” jump is incorrect
### System Information Name: opencv-python Version: 4.8.0.76 System: Ubuntu 20.04 Python Version: 3.11.9 ### Detailed description I use cap.set(cv2.CAP_PROP_POS_FRAMES, start_frame) to jump to the 360th frame, with a frame rate of 60, which is 6 seconds, but I found that the frame after the jump is the video frame at the 12th second. I have only found a problem in one video,Other video frame rates are all 30 or 29, and there is no problem in other videos. What could be the reason? Below is my code: cap = cv2.VideoCapture(video_path) if not cap.isOpened(): print(f"Error: Unable to open video file {video_path}") return # 获取视频的FPS和总帧数 fps = cap.get(cv2.CAP_PROP_FPS) total_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) # 计算从第几帧开始 start_frame = int(fps * start_time) end_frame = int(fps * end_time) # 跳转到起始帧 cap.set(cv2.CAP_PROP_POS_FRAMES, start_frame) frame_interval = round(fps / fps_target) # 每秒需要提取的帧数(15帧每秒) ### Steps to reproduce Video privacy concerns make it inconvenient to upload, sorry ### Issue submission checklist - [x] I report the issue, it's not a question - [x] I checked the problem with documentation, FAQ, open issues, forum.opencv.org, Stack Overflow, etc and have not found any solution - [ ] I updated to the latest OpenCV version and the issue is still there - [x] There is reproducer code and related data files (videos, images, onnx, etc)
bug,category: videoio,incomplete
low
Critical
2,806,216,788
storybook
[Storybook 9]: Remove deprecations
The following deprecations will be removed in Storybook 9: | Feature | Description | File Path | |---------|-------------|-----------| | Addon Backgrounds | features.backgroundsStoryGlobals | code/addons/backgrounds/src/preview.ts | | Addon Viewport | features.viewportStoryGlobals | code/addons/viewport/src/preview.ts | | Addon Viewport | style function type | code/addons/viewport/src/types.ts | | Addon_TestProviderType | title, description | code/core/src/types/modules/addons.ts | | Addon_TypesEnum | experimental_SIDEBAR_BOTTOM, experimental_SIDEBAR_TOP | code/core/src/types/modules/addons.ts | | API_Layout | isToolshown | code/core/src/types/modules/api.ts | | API_Notification | DeprecatedIconType | code/core/src/types/modules/api.ts | | API_Provider | serverChannel | code/core/src/types/modules/api.ts | | CLI | renderer svelte template | code/core/src/cli/helpers.ts | | Codemod | the entire package -> move to automigrations | code/lib/codemod | | component | ButtonProps (isLink, primary, secondary, tertiary, gray, inForm, small, outline, containsIcon) | code/core/src/components/components/Button/Button.tsx | | component | IconButtonSkeleton | code/core/src/components/components/bar/button.tsx | | component | Icons | code/core/src/components/components/icon/icon.tsx | | component | Symbols | code/core/src/components/components/icon/icon.tsx | | CSFFile | _fileName, _makeTitle | code/core/src/csf-tools/CsfFile.ts | | DocsOptions | autodocs | code/core/src/types/modules/core-common.ts | | event | TESTING_MODULE_RUN_ALL_REQUEST | code/core/src/core-events/index.ts | | Lib Test | fn | code/lib/test/src/spy.ts | | Lib Test | Mock | code/lib/test/src/spy.ts | | Manager API | DeprecatedState | code/core/src/manager-api/root.tsx | | Manager API | getAddonState | code/core/src/manager-api/modules/addons.ts | | Manager API | setAddonState | code/core/src/manager-api/modules/addons.ts | | parameters.a11y.manual | | | | Portable Stories | Legacy Portable Stories format | code/renderers/react/src/__test__/portable-stories-legacy.test.tsx | | Portable Stories | Consolidate this renderContext with Context in SB 9.0 | code/core/src/preview-api/modules/store/csf/portable-stories.ts | | Portable Stories | globals (replaced by initialGlobals) | code/core/src/preview-api/modules/store/csf/portable-stories.ts | | Preview API | addons, mockChannel | code/core/src/preview-api/index.ts | | Preview API | fromId | code/core/src/preview-api/modules/store/StoryStore.ts | | Preview API | getSetStoriesPayload | code/core/src/preview-api/modules/store/StoryStore.ts | | Preview API | getStoriesJsonData | code/core/src/preview-api/modules/store/StoryStore.ts | | Preview API | globals (replaced by initialGlobals) | code/core/src/preview-api/modules/store/csf/normalizeProjectAnnotations.ts | | Preview API | raw | code/core/src/preview-api/modules/store/StoryStore.ts | | Preview API | serverChannel | code/core/src/preview-api/modules/preview-web/Preview.tsx | | Preview API | storyStore | code/core/src/preview-api/modules/preview-web/Preview.tsx | | project annotations | renderToDOM | code/core/src/types/modules/story.ts | | type | Addon_SidebarBottomType | code/core/src/types/modules/addons.ts | | type | Addon_SidebarTopType | code/core/src/types/modules/addons.ts | | type | frameworkToRenderer | code/core/src/cli/helpers.ts | | type | subcomponents | code/core/src/types/modules/addons.ts | | type | SupportedRenderers | code/core/src/cli/project_types.ts |
bug,needs triage
low
Minor
2,806,238,913
pytorch
[inductor][torchbench] Unsupported operator issue when running the torch_multimodal_clip model with batch size 4.
### 🐛 Describe the bug I'm conducting an accuracy test for the `torch_multimodal_clip `model. I noticed that the graph breaks count is 0 when using a batch size of 1. However, when increasing the batch size to 4, the graph breaks count rises to 3. Upon printing the graph break details, I discovered the unsupported operator issue occurring with batch size 4. ```shell python benchmarks/dynamo/torchbench.py --accuracy --amp -dcpu --inference -n50 --inductor --timeout 9000 --dynamic-shapes --dynamic-batch-only --freezing --only torch_multimodal_clip --print-graph-breaks --explain --batch-size 4 ``` ``` loading model: 0it [00:02, ?it/s] cpu eval torch_multimodal_clip WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] Graph break in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313 V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix) V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] User code traceback: V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/benchmarks/dynamo/common.py", line 2782, in run_n_iterations V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] self.model_iter_fn(mod, inputs, collect_outputs=False) V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/benchmarks/dynamo/torchbench.py", line 444, in forward_pass V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] return mod(*inputs) V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/opt/conda/lib/python3.10/site-packages/torchmultimodal/models/clip/model.py", line 71, in forward V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] embeddings_a = self.encoder_a(features_a) V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/opt/conda/lib/python3.10/site-packages/torchmultimodal/models/clip/image_encoder.py", line 108, in forward V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = self.encoder(x) V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 517, in forward V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] output = mod( V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 913, in forward V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = x + self._sa_block( V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 934, in _sa_block V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = self.self_attn( V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/activation.py", line 1313, in forward V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] return torch._native_multi_head_attention( V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] V0123 16:00:42.211000 10589 torch/_dynamo/symbolic_convert.py:467] [1/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313 V0123 16:00:42.211000 10589 torch/_dynamo/symbolic_convert.py:467] [1/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix) V0123 16:00:42.466000 10589 torch/_dynamo/symbolic_convert.py:467] [2/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313 V0123 16:00:42.466000 10589 torch/_dynamo/symbolic_convert.py:467] [2/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix) V0123 16:00:42.569000 10589 torch/_dynamo/symbolic_convert.py:467] [3/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313 V0123 16:00:42.569000 10589 torch/_dynamo/symbolic_convert.py:467] [3/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix) pass WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] Dynamo produced 6 graphs covering 209 ops with 3 graph breaks (1 unique) ``` For reference, below is the log of batch size 1 ```shell python benchmarks/dynamo/torchbench.py --accuracy --amp -dcpu --inference -n50 --inductor --timeout 9000 --dynamic-shapes --dynamic-batch-only --freezing --only torch_multimodal_clip --print-graph-breaks --explain --batch-size 1 ``` ``` loading model: 0it [00:02, ?it/s] cpu eval torch_multimodal_clip WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] pass WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] Dynamo produced 1 graphs covering 744 ops with 0 graph breaks (0 unique) ``` ### Versions PyTorch version: 2.7.0a0+git9ae35b8 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.26.4 Libc version: glibc-2.35 Python version: 3.10.16 | packaged by conda-forge | (main, Dec 5 2024, 14:16:10) [GCC 13.3.0] (64-bit runtime) Python platform: Linux-5.15.0-130-generic-x86_64-with-glibc2.35 Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 52 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 384 On-line CPU(s) list: 0-383 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) 6972P CPU family: 6 Model: 173 Thread(s) per core: 2 Core(s) per socket: 96 Socket(s): 2 Stepping: 1 CPU max MHz: 3900.0000 CPU min MHz: 800.0000 BogoMIPS: 4800.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 9 MiB (192 instances) L1i cache: 12 MiB (192 instances) L2 cache: 384 MiB (192 instances) L3 cache: 960 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-95,192-287 NUMA node1 CPU(s): 96-191,288-383 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS Not affected; BHI BHI_DIS_S Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] bert_pytorch==0.0.1a4 [pip3] functorch==1.14.0a0+b71aa0b [pip3] mypy==1.14.1 [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.26.4 [pip3] onnx==1.17.0 [pip3] torch==2.7.0a0+git9ae35b8 [pip3] torch_geometric==2.4.0 [pip3] torchaudio==2.6.0a0+b6d4675 [pip3] torchdata==0.7.0a0+11bb5b8 [pip3] torchmultimodal==0.1.0b0 [pip3] torchtext==0.16.0a0+b0ebddc [pip3] torchvision==0.19.0a0+d23a6e1 [conda] bert-pytorch 0.0.1a4 dev_0 <develop> [conda] functorch 1.14.0a0+b71aa0b pypi_0 pypi [conda] mkl 2025.0.0 h901ac74_941 conda-forge [conda] mkl-include 2025.0.0 hf2ce2f3_941 conda-forge [conda] numpy 1.26.4 pypi_0 pypi [conda] torch 2.7.0a0+git9ae35b8 dev_0 <develop> [conda] torch-geometric 2.4.0 pypi_0 pypi [conda] torchaudio 2.6.0a0+b6d4675 pypi_0 pypi [conda] torchdata 0.7.0a0+11bb5b8 pypi_0 pypi [conda] torchmultimodal 0.1.0b0 pypi_0 pypi [conda] torchtext 0.16.0a0+b0ebddc pypi_0 pypi [conda] torchvision 0.19.0a0+d23a6e1 pypi_0 pypi cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @chauhang @penguinwu @cpuhrsch @bhosmer @drisspg @soulitzer @davidberard98 @YuqingJ
module: nn,triaged,oncall: pt2,module: sdpa
low
Critical
2,806,242,455
vscode
Source control not working after adding turborepo to project
Does this issue occur when all extensions are disabled?: Yes - VS Code Version: - Version: 1.96.1 (Universal) - Commit: 42b266171e51a016313f47d0c48aca9295b9cbb2 - Date: 2024-12-17T17:50:05.206Z (1 mo ago) - Electron: 32.2.6 - ElectronBuildId: 10629634 - Chromium: 128.0.6613.186 - Node.js: 20.18.1 - V8: 12.8.374.38-electron.0 - OS: Darwin arm64 23.5.0 - OS and Version: MacOs sonoma 14.5 ### Bug Summary and Steps to Reproduce Bug Summary: After we added turborepo to our project, source control stopped working properly. After making a change to any file if I go to gitlens and select the file then I see "The editor could not be opened because the file was not found." clicking on try again button I see : "Unable to read file 'git:/Users/nikunjgoyal/work/app/apps/web/Dockerfile?{"path":"/Users/nikunjgoyal/work/app/apps/web/Dockerfile","ref":""}' (Error: Unable to resolve nonexistent file 'git:/Users/nikunjgoyal/work/app/apps/web/Dockerfile?{"path":"/Users/nikunjgoyal/work/app/apps/web/Dockerfile","ref":""}')" ![Image](https://github.com/user-attachments/assets/a1a30503-d9b3-45e0-b590-663ead27106e) Steps to reproduce: 1. Not sure how this can be reproduced Expected behavior: I should be able to compare differences in source control
info-needed,git
low
Critical
2,806,249,840
vscode
Git error dialog swap button functions
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> When an error occurs from the Git hooks, a dialog appears with buttons: - Open Git Log - Show Command Output - Cancel The problem is that the button click events are swapped. Cancel works correctly, but the other 2 buttons are wrong. - Open Git Log -> Shows the command output - Show Command Output -> Opens the Git Log The solution is simple: Either swap the button title or the button click events. But I'm not sure what component is showing this dialog. Either VSCode or [Husky](https://github.com/typicode/husky)? Background: I'm working on an Angular project using the Husky node module for Git hooks. The dialog appears when a hook task failed (e.g. linter detected a problem). ![Image](https://github.com/user-attachments/assets/1c9ba625-4d7f-47cb-8ed2-acc6c7b6c60d) Version: 1.96.4 (user setup) Electron: 32.2.6 Chromium: 128.0.6613.186 Node.js: 20.18.1 V8: 12.8.374.38-electron.0 OS: Windows_NT x64 10.0.26100
info-needed,git
low
Critical
2,806,250,364
storybook
[Storybook 9]: Drop support for tooling/integrations/frameworks
### Describe the bug The official support and maintenance of the following third-party toolings/integrations/frameworks will be removed in Storybook 9: | Category | Tool | Versions | |----------|------|----------| | JavaScript Engine | Node.js | v18 | | Package Manager | npm | v8, v9 | | Package Manager | yarn | v3 | | Package Manager | pnpm | v7, v8 | | Renderer | React | v16, v17 (up for discussion) | | Renderer | Angular | v15, v16, v17 | | Renderer | Lit | v1, v2 (up for discussion) | | Builders | Vite | v4 | | Language | Typescript | < v4.9 (4.9 introduces satisfies support) | | Renderer | Svelte | < v4 |
bug,needs triage
low
Critical
2,806,254,072
electron
Calls to window.open in offscreen documents do not work
### Preflight Checklist - [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success. ### Electron Version 34.0.1 ### What operating system(s) are you using? Other Linux ### Operating System Version 6.12.6-1 ### What arch are you using? x64 ### Last Known Working Electron version _No response_ ### Expected Behavior `window.open` calls in offscreen documents created with `chrome.offscreen.createDocument` routes the window open request to the webcontents window open handler set with `setWindowOpenHandler` or just opens the window in a new `BrowserWindow` and returns the `Window` object. ### Actual Behavior `window.open` does nothing and returns null. ### Testcase Gist URL https://github.com/BrettCleary/electron_bug_example ### Additional Information Please see https://github.com/BrettCleary/electron_bug_example for a simple, reproducible example. I don't see an option to add an extension folder in electron fiddle and there is too much setup for a gist. Also I know that full extension support is not a goal of core electron developers, but given that `chrome.offscreen.createDocument` works, I would expect `window.open` to work in the offscreen doc. If you could even point me in the right direction to start troubleshooting, I could try to open a PR with a fix as well 🙏. I would love the opportunity to help with core electron development 🚀.
platform/linux,bug :beetle:,34-x-y
low
Critical
2,806,255,097
rust
Long type name emitted if the type name is part of a suggestion
### Code ```Rust use diesel::prelude::*; //diesel = { version = "2.2.6", default-features = false, features = ["postgres"] } diesel::table! { foo (id) { id -> Int4, name -> Text, } } fn main() { diesel::debug_query( &foo::table .filter(foo::id.eq(42)) .order(foo::name.desc()) .limit(10) .offset(42), ); } ``` ### Current output ```Shell error[E0282]: type annotations needed --> src/main.rs:11:5 | 11 | diesel::debug_query( | ^^^^^^^^^^^^^^^^^^^ cannot infer type of the type parameter `DB` declared on the function `debug_query` | help: consider specifying the generic arguments | 11 | diesel::debug_query::<DB, SelectStatement<FromClause<table>, query_builder::select_clause::DefaultSelectClause<FromClause<table>>, query_builder::distinct_clause::NoDistinctClause, query_builder::where_clause::WhereClause<diesel::expression::grouped::Grouped<diesel::expression::operators::Eq<columns::id, diesel::expression::bound::Bound<Integer, i32>>>>, query_builder::order_clause::OrderClause<diesel::expression::operators::Desc<columns::name>>, LimitOffsetClause<LimitClause<diesel::expression::bound::Bound<BigInt, i64>>, OffsetClause<diesel::expression::bound::Bound<BigInt, i64>>>>>( | ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ ``` ### Desired output ```Shell error[E0282]: type annotations needed --> src/main.rs:11:5 | 11 | diesel::debug_query( | ^^^^^^^^^^^^^^^^^^^ cannot infer type of the type parameter `DB` declared on the function `debug_query` | help: consider specifying the generic arguments | 11 | diesel::debug_query::<DB, _>( | ++ ``` ### Rationale and extra context Emitting the query type there is not helpful for the users as it is long, makes the error message harder to read and ultimately is not even required to fix the code snippet. Fixing the code snippet only requires to specify the `DB` generic type there. Ideally rustc would be able to suggest possible types for `DB` there, given the existing constraint on `debug_query`. In the example above that's even only satisfied by one type (`diesel::pg::Pg`). ### Other cases If you change the exact query that you pass to `debug_query` the output changes. You can make the query more complex and therefore get a much more complex and longer query type in the compiler output. ### Rust Version ```Shell rustc 1.86.0-nightly (649b995a9 2025-01-22) binary: rustc commit-hash: 649b995a9febd658b2570160703dff6fdc038ab2 commit-date: 2025-01-22 host: x86_64-unknown-linux-gnu release: 1.86.0-nightly LLVM version: 19.1.7 ``` ### Anything else? cc @estebank as they expressed interested in such cases
A-diagnostics,T-compiler
low
Critical
2,806,269,401
flutter
[BUG] Error: PathNotFoundException during symbolic link resolution on Linux (/proc/self/exe not found)
### Steps to reproduce 1. Run the Flutter application on a Linux device. 2. Observe the error during startup. ### Expected results The application should start without errors, and the `path_provider_linux` package should successfully resolve symbolic links for the application support path. ### Actual results `PathNotFoundException: Cannot resolve symbolic links, path = '/proc/self/exe' (OS Error: No such file or directory, errno = 2)` ### Code sample <details open><summary>Code sample</summary> ```dart Future<void> main() async { WidgetsFlutterBinding.ensureInitialized(); await init(); // Injection container initialization runApp(MyApp()); } init(){ final prefsWithCache = await SharedPreferencesWithCache.create( cacheOptions: const SharedPreferencesWithCacheOptions(), ); /// Register AppPreferences and its implementation di.registerLazySingleton<AppPreferences>( () => AppPreferencesImpl( prefsWithCache, ), ); } class AppPreferencesImpl implements AppPreferences { late SharedPreferencesWithCache _sharedPreferences; set sharedPreferences(SharedPreferencesWithCache value) { _sharedPreferences = value; } AppPreferencesImpl(SharedPreferencesWithCache sharedPrefs) { _sharedPreferences = sharedPrefs; initSharedPrefs(); } initSharedPrefs() async { _sharedPreferences = await SharedPreferencesWithCache.create( cacheOptions: const SharedPreferencesWithCacheOptions(), ); } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> ![Image](https://github.com/user-attachments/assets/8558bdb7-9632-4095-8cc5-7bb5855c4680) </details> ### Logs <details open><summary>Logs</summary> ```console (innogec_mobile:161765): GLib-GObject-WARNING **: 09:28:46.135: g_object_weak_unref: couldn't find weak ref 0x75a497ff0b50(0x29c234d8) Restarted application in 684ms. [ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PathNotFoundException: Cannot resolve symbolic links, path = '/proc/self/exe' (OS Error: No such file or directory, errno = 2) #0 _checkForErrorResponse (dart:io/common.dart:56:9) common.dart:56 #1 FileSystemEntity.resolveSymbolicLinks.<anonymous closure> (dart:io/file_system_entity.dart:318:7) file_system_entity.dart:318 <asynchronous suspension> #2 PathProviderLinux._getExecutableName (package:path_provider_linux/src/path_provider_linux.dart:92:9) path_provider_linux.dart:92 <asynchronous suspension> #3 PathProviderLinux._getId (package:path_provider_linux/src/path_provider_linux.dart:100:30) path_provider_linux.dart:100 <asynchronous suspension> #4 PathProviderLinux.getApplicationSupportPath (package:path_provider_linux/src/path_provider_linux.dart:51:48) path_provider_linux.dart:51 <asynchronous suspension> #5 _getLocalDataFile (package:shared_preferences_linux/shared_preferences_linux.dart:339:29) shared_preferences_linux.dart:339 <asynchronous suspension> #6 _reload (package:shared_preferences_linux/shared_preferences_linux.dart:354:31) shared_preferences_linux.dart:354 <asynchronous suspension> #7 SharedPreferencesAsyncLinux._readPreferences (package:shared_preferences_linux/shared_preferences_linux.dart:323:28) shared_preferences_linux.dart:323 <asynchronous suspension> #8 SharedPreferencesAsyncLinux._readAll (package:shared_preferences_linux/shared_preferences_linux.dart:300:34) shared_preferences_linux.dart:300 <asynchronous suspension> #9 SharedPreferencesWithCache.reloadCache (package:shared_preferences/src/shared_preferences_async.dart:225:9) shared_preferences_async.dart:225 <asynchronous suspension> #10 SharedPreferencesWithCache.create (package:shared_preferences/src/shared_preferences_async.dart:200:5) shared_preferences_async.dart:200 <asynchronous suspension> #11 init (package:innogec_mobile/injection_container.dart:58:26) injection_container.dart:58 <asynchronous suspension> #12 mainCommon (package:innogec_mobile/main_common.dart:16:3) main_common.dart:16 <asynchronous suspension> ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [✓] Android toolchain - develop for Android devices (Android SDK version 35.0.0) • Android SDK at /home/user/Android/Sdk • Platform android-35, build-tools 35.0.0 • Java binary at: /snap/android-studio/161/jbr/bin/java • Java version OpenJDK Runtime Environment (build 17.0.10+0-17.0.10b1087.21-11609105) • All Android licenses accepted. [✓] Chrome - develop for the web • Chrome at google-chrome [✓] Linux toolchain - develop for Linux desktop • clang version 10.0.0-4ubuntu1 • cmake version 3.16.3 • ninja version 1.10.0 • pkg-config version 0.29.1 [✓] Android Studio (version 2023.3) • Android Studio at /snap/android-studio/157 • Flutter plugin version 79.0.2 • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.10+0-17.0.10b1087.21-11572160) [✓] Android Studio (version 2024.1) • Android Studio at /snap/android-studio/161 • Flutter plugin version 83.0.2 • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.10+0-17.0.10b1087.21-11609105) [✓] IntelliJ IDEA Ultimate Edition (version 2024.3) • IntelliJ at /snap/intellij-idea-ultimate/561 • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart [✓] VS Code (version 1.96.4) • VS Code at /snap/code/current/usr/share/code • Flutter extension version 3.102.0 [✓] Connected device (2 available) • Linux (desktop) • linux • linux-x64 • Ubuntu 22.04.5 LTS 6.8.0-51-generic • Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83 [✓] Network resources • All expected network resources are available. • No issues found! ``` </details>
waiting for customer response,in triage
low
Critical
2,806,272,957
vscode
Crashing when clicking on File menu with native title bar
Type: <b>Bug</b> Steps to Reproduce: 1. Arch linux with wayland (hyprland compositor) 2. Hover/click the file menu 2. All windows crash This looks very similar to the issue discussed at https://github.com/microsoft/vscode/issues/230604 But at least for me, I'm on latest versions and made sure that the recommended workaroudn of removing meta/super key from the keybind is applied, still - vscode crashes for me. Still happens if I launch with `code --disable-extensions` If I downgrade to 1.94.0 it no longer crashes but I cannot use extensions such as GH copilot. VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Linux x64 6.12.10-arch1-1 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) Ultra 7 165H (22 x 3804)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off<br>webnn: disabled_off| |Load (avg)|4, 2, 2| |Memory (System)|62.24GB (51.33GB free)| |Process Argv|-enable-features=UseOzonePlatform,WaylandWindowDecorations --ozone-platform=wayland --crash-reporter-id 83481648-87d4-4456-9caf-70c8cdfed42f| |Screen Reader|no| |VM|0%| |DESKTOP_SESSION|undefined| |XDG_CURRENT_DESKTOP|Hyprland| |XDG_SESSION_DESKTOP|Hyprland| |XDG_SESSION_TYPE|wayland| </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 vscod805:30301674 binariesv615:30325510 vsaa593:30376534 py29gd2263:31024239 vscaat:30438848 c4g48928:30535728 azure-dev_surveyone:30548225 a9j8j154:30646983 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 2e7ec940:31000449 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupytercf:31046870 nativerepl2:31139839 pythonrstrctxt:31112756 nativeloc2:31192216 cf971741:31144450 iacca1:31171482 notype1cf:31157160 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 6074i472:31201624 dwoutputs:31217127 hdaa2157:31222309 copilot_t_ci:31222730 ``` </details> <!-- generated by issue reporter -->
bug,upstream,freeze-slow-crash-leak,upstream-issue-linked,chromium,mitigated
low
Critical
2,806,284,064
transformers
resume_from_checkpoint failed when using PEFT LORA
### System Info - `transformers` version: 4.43.1 - Platform: Linux-6.5.0-45-generic-x86_64-with-glibc2.35 - Python version: 3.10.14 - Huggingface_hub version: 0.24.7 - Safetensors version: 0.4.5 - Accelerate version: 1.1.0 - Accelerate config: - compute_environment: LOCAL_MACHINE - distributed_type: MULTI_GPU - mixed_precision: no - use_cpu: False - debug: False - num_processes: 2 - machine_rank: 0 - num_machines: 1 - gpu_ids: all - rdzv_backend: static - same_network: True - main_training_function: main - enable_cpu_affinity: False - downcast_bf16: no - tpu_use_cluster: False - tpu_use_sudo: False - tpu_env: [] - PyTorch version (GPU?): 2.4.1+cu124 (True) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: <fill in> - Using GPU in script?: <fill in> - GPU type: NVIDIA H100 NVL ### Who can help? @muellerzr @SunMarc ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction I am training my own model (a vision-language model) with LoRA using the Trainer. However, when I try to use resume_from_checkpoint, I encounter the following error: ``` Traceback (most recent call last): File "/dataT0/Free/hengyue/dataF0_save/new/LMM/training_Trainer.py", line 137, in <module> trainer.train(resume_from_checkpoint=True) File "/dataF0/Free/hengyue/miniconda3/envs/doraemon24/lib/python3.10/site-packages/transformers/trainer.py", line 2143, in train self._load_from_checkpoint(resume_from_checkpoint) File "/dataF0/Free/hengyue/miniconda3/envs/doraemon24/lib/python3.10/site-packages/transformers/trainer.py", line 2851, in _load_from_checkpoint if len(active_adapters) > 1: TypeError: object of type 'method' has no len() ``` My Code Shows below: ``` lora_config = LoraConfig( task_type=TaskType.CAUSAL_LM, inference_mode=False, r=8, lora_alpha=32, lora_dropout=0.05, target_modules=["q_proj", "k_proj", "v_proj", "o_proj"], ) model = MyModel() model = get_peft_model(model, lora_config) model.print_trainable_parameters() device = torch.device("cuda" if torch.cuda.is_available() else "cpu") train_dataset = Mydataset() val_dataset = Valdataset() training_args = TrainingArguments( output_dir="./results", evaluation_strategy="epoch", save_strategy="epoch", learning_rate=1e-4, per_device_train_batch_size=2, per_device_eval_batch_size=1, num_train_epochs=11, weight_decay=0.01, logging_dir="./logs", logging_steps=10, save_total_limit=2, report_to="wandb", dataloader_num_workers = 4, ) trainer = Trainer( model=model, args=training_args, train_dataset=train_dataset, eval_dataset=val_dataset, data_collator=custom_collate_fn, ) trainer.train(resume_from_checkpoint=True) trainer.save_state() trainer.save_model("./test_model") ``` ### Expected behavior Expected the model should continue training from the checkpoint.
bug
low
Critical
2,806,293,420
react
[DevTools Bug]: Support Iterables in React DevTools
### Website or app React DevTools ### Repro steps 1.Open a React app using React 19.0.0. 2.Pass an Iterable<T> to a component's props or state. 3.Inspect the component using React DevTools. 4.Observe the display or handling of the Iterable. ### How often does this bug happen? Every time ### DevTools package (automated) _No response_ ### DevTools version (automated) _No response_ ### Error message (automated) _No response_ ### Error call stack (automated) ```text ``` ### Error component stack (automated) ```text ``` ### GitHub query string (automated) ```text ```
Type: Bug,Status: Unconfirmed,Component: Developer Tools
medium
Critical
2,806,304,172
terminal
The lowercase I glyph (U+0069) in JuliaMono renders incorrectly
### Windows Terminal version 1.22.241118002-preview, also happens on 1.21.3231.0 ### Windows build number 10.0.26100.2894 ### Other Software _No response_ ### Steps to reproduce 1. Download the JuliaMono font if you haven't yet. 2. Set the terminal font to JuliaMono. 3. Open a terminal and type in "i". The glyph renders incorrectly for me. ### Expected Behavior I was expecting this: ![Image](https://github.com/user-attachments/assets/abc70cee-b149-4586-9e1e-923ead8d39c9) Note the serifs in the lowercase I. ### Actual Behavior I got this behavior: ![Image](https://github.com/user-attachments/assets/e8f682d0-f3bc-4792-becf-2bbcbad68d4c) Some notes: - Other variations of the lowercase I (î, ï, ı etc.) render with serifs. - Uppercase I renders with serifs, as seen in the picture.
Issue-Bug,Needs-Triage
low
Major
2,806,313,473
react
Clarify Behavior of useSyncExternalStore When Using Server-Side Rendering (SSR)
# Title: Clarify Behavior of useSyncExternalStore When Using Server-Side Rendering (SSR) ## Description: After reviewing PR #29760, which addresses the optimization of useSyncExternalStore in React, I noticed that the behavior of this hook when used in server-side rendering (SSR) scenarios isn't clearly documented. Specifically, the following points need clarification: How should developers manage subscriptions during SSR? What best practices should be followed to avoid potential inconsistencies when hydrating the client-side store? <!-- Please provide a clear and concise description of what the bug is. Include screenshots if needed. Please test using the latest version of the relevant React packages to make sure your issue has not already been fixed. --> React version: 19 ## Steps To Reproduce 1. Set up the environment: Include details about the environment where the issue occurs (e.g., React version, Node.js version, browser, etc.). 2. Create a minimal project: Describe how to create a minimal reproducible example that isolates the bug (e.g., a simple React app using create-react-app). 3. Write relevant code: Provide the key parts of the code where the issue happens (e.g., using a specific React feature, hook, or component). 4. Run the project: Explain how to execute the project to trigger the bug. 5. Observe the behavior: Describe what happens when the steps are followed (the actual behavior), compared to what you expected to happen. <!-- Your bug will get fixed much faster if we can run your code and it doesn't have dependencies other than React. Issues without reproduction steps or code examples may be immediately closed as not actionable. --> Link to code example: <!-- Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a repository on GitHub, or provide a minimal code example that reproduces the problem. You may provide a screenshot of the application if you think it is relevant to your bug report. Here are some tips for providing a minimal example: https://stackoverflow.com/help/mcve. --> Please use one of the following techniques to produce a simple, replicable sample of the bug: Give a link to a CodeSandbox project that illustrates the problem. GitHub Repository: Provide a link to a small project on GitHub that replicates the issue. Inline Code: Use appropriate formatting to include the code right in the bug report if it is brief and simple. ## The current behavior There isn't much advice on SSR-specific circumstances in the present material. ## The expected behavior The documentation should explicitly detail how useSyncExternalStore interacts with SSR. Provide examples of correct usage in SSR and hydration contexts. ## References - Related Pull Request: [29760](https://github.com/facebook/react/pull/29760)
Status: Unconfirmed
medium
Critical
2,806,323,057
godot
There is no order in calling method _input_event() of Area2D. Expectation is order by node position in sence tree
### Tested versions 4.3-stable ### System information MacOS 15 ### Issue description See the picture below, i have created a minimal reproduction project following. This issue make it impossible to control the propagation order of input event. ![Image](https://github.com/user-attachments/assets/598e5bfe-e9a5-476f-8d6f-7ad74edc9f4e) ### Steps to reproduce NA ### Minimal reproduction project (MRP) [buttest.zip](https://github.com/user-attachments/files/18517822/buttest.zip)
discussion,documentation,topic:physics
low
Minor
2,806,325,434
bitcoin
test: intermittent issue in p2p_1p1c_network.py
Looks like this still happens: https://github.com/bitcoin/bitcoin/pull/31713#issuecomment-2609212276 So this is likely a non-timeout issue. Note that the test has immediate tx-relay (`self.noban_tx_relay`) enabled, so trickle delay should not be an issue. _Originally posted by @maflcko in https://github.com/bitcoin/bitcoin/issues/31701#issuecomment-2609226329_
CI failed
low
Major
2,806,330,642
vscode
VSCode will rollback file content
Type: <b>Bug</b> I am using vs2022 and vscode for same folder, but vscode will rollback other editor's change. First use other editor change the file, save, then use vs code open it, the file will be rollback, I tested , the reason is vs code will use cache "C:\Users\qwe\AppData\Roaming\Code", if delete this folder, it will fix, but vs code do not provide a option to disable the "cache", Please fix the bug, or there is another way to fix this please tell me , thank you! VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z) OS version: Windows_NT x64 10.0.19045 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) N100 (4 x 806)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|undefined| |Memory (System)|15.68GB (3.70GB free)| |Process Argv|--crash-reporter-id 55627040-681b-4624-ae9b-2941d63e8d17| |Screen Reader|no| |VM|40%| </details><details><summary>Extensions (9)</summary> Extension|Author (truncated)|Version ---|---|--- copilot|Git|1.258.0 copilot-chat|Git|0.23.2 better-cpp-syntax|jef|1.27.1 marscode-extension|Mar|1.1.58 cmake-tools|ms-|1.19.52 cpptools|ms-|1.22.11 cpptools-extension-pack|ms-|1.3.0 BuildOutputColorizer|Ste|0.1.8 cmake|twx|0.0.17 (1 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 vscod805:30301674 binariesv615:30325510 vsaa593cf:30376535 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 h48ei257:31000450 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupytercf:31046870 2f103344:31071589 nativerepl1:31139838 pythonrstrctxt:31112756 nativeloc2:31192216 cf971741:31144450 iacca1:31171482 notype1:31157159 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 6074i472:31201624 dwoutputs:31217127 9064b325:31222308 copilot_t_ci:31222730 ``` </details> <!-- generated by issue reporter -->
workbench-copilot
low
Critical
2,806,376,723
rust
Cleanup "no panic support" tests
There are a bunch of tests that say `ignore-wasm` or `ignore-emscripten` or `ignore-$platform` with a comment "no panic support" or "no panic". It would be good to audit them and see if one or more of the following capability-based conditional test execution directives can be used instead: - `//@ needs-unwind` - `//@ needs-threads` - `//@ needs-subprocess`[^1] If those are not sufficient, it would be good to improve the ignore reason anyway. [^1]: I'm still working on implementing this directive over at #128295.
C-cleanup,A-testsuite,T-compiler,E-medium,A-panic
low
Minor
2,806,379,010
flutter
FFI: SIGSEGV crash in release mode when handling image data between C++ and Dart, woks in debug mode
### Steps to reproduce Clone repository: https://github.com/rmatif/Local-Diffusion Build and run in debug mode - works correctly Build release APK: flutter build apk --release --target-platform android-arm64 Install and run release APK Initialize model and generate an image Observe crash after image generation completes Already attempted fixes: - Added ProGuard rules to preserve FFI classes <details open><summary>proguard-rules.pro</summary> ```pro -keep class * extends ffi.Struct { *; } -keep class * implements ffi.Finalizable { *; } -keep class ffi_bindings.FFIBindings { # Keep the FFIBindings class itself *; # Keep all members } -keep class sd_image.SDImage { # Keep the SDImage class *; # Keep all members } -keep class com.example.local_diffusion.** { # Keep your main app package. Be more specific if possible. *; } # If you have specific callback functions that are only called from native code: -keepclassmembers class ffi_bindings.FFIBindings { public static void _staticLogCallback(int, ffi.Pointer, ffi.Pointer); # Example, adjust names public static void _staticProgressCallback(int, int, double, ffi.Pointer); # Example, adjust names } -keepclassmembers class sd_image.SDImage { # Keep members of SDImage public int width; # Keep specific fields accessed from native side (if any direct access) public int height; public int channel; public ffi.Pointer data; *; # Keep all members for safety, refine if needed later } -keepclassmembers class com.example.local_diffusion.stable_diffusion_processor.StableDiffusionProcessor { # Keep members if needed *; # Keep all members for safety, refine if needed later } -keep,allowshrinking,allowobfuscation class com.example.local_diffusion.** { # Keep your main app package <methods>; <fields>; <init>(...); } -keepattributes Signature -keepattributes *Annotation* -keep class dart.ffi.* { *; } -keep class io.flutter.plugin.platform.* { *; } ``` </details> - Disabled minification and resource shrinking <details open><summary>build.gradle</summary> ```gradle plugins { id "com.android.application" id "kotlin-android" id "dev.flutter.flutter-gradle-plugin" } def localProperties = new Properties() def localPropertiesFile = rootProject.file('local.properties') if (localPropertiesFile.exists()) { localPropertiesFile.withReader('UTF-8') { reader -> localProperties.load(reader) } } def flutterVersionCode = localProperties.getProperty('flutter.versionCode') if (flutterVersionCode == null) { flutterVersionCode = '1' } def flutterVersionName = localProperties.getProperty('flutter.versionName') if (flutterVersionName == null) { flutterVersionName = '1.0' } android { namespace "com.example.local_diffusion" compileSdkVersion flutter.compileSdkVersion ndkVersion flutter.ndkVersion compileOptions { sourceCompatibility JavaVersion.VERSION_1_8 targetCompatibility JavaVersion.VERSION_1_8 } kotlinOptions { jvmTarget = '1.8' } sourceSets { main.java.srcDirs += 'src/main/kotlin' } defaultConfig { // TODO: Specify your own unique Application ID (https://developer.android.com/studio/build/application-id.html). applicationId "com.example.local_diffusion" // You can update the following values to match your application needs. // For more information, see: https://docs.flutter.dev/deployment/android#reviewing-the-gradle-build-configuration. minSdkVersion flutter.minSdkVersion targetSdkVersion flutter.targetSdkVersion versionCode flutterVersionCode.toInteger() versionName flutterVersionName } buildTypes { release { // TODO: Add your own signing config for the release build. // Signing with the debug keys for now, so `flutter run --release` works. minifyEnabled false shrinkResources false signingConfig signingConfigs.debug proguardFiles getDefaultProguardFile('proguard-android.txt'), 'proguard-rules.pro' } } } flutter { source '../..' } dependencies {} ``` </details> - Modified memory management approach - The issue appears specific to release mode builds while handling FFI image data transfer ### Expected results Image should be displayed and saved successfully after generation, same as in debug mode. ### Actual results App crashes with SIGSEGV (signal 11) after image generation completes `signal 11 (SIGSEGV), code 2 (SEGV_ACCERR), fault addr 0x0000007600aac080 pid: 17513, tid: 18005, name: DartWorker ` ### Code sample <details open><summary>Code sample</summary> ```dart if (result.address != 0) { final image = result.cast<SDImage>().ref; final width = (message['width'] as num).toInt(); final height = (message['height'] as num).toInt(); final channels = image.channel; final totalBytes = width * height * channels; final rgbaBytes = Uint8List(width * height * 4); // Crash occurs during image data processing } ``` </details> ### Screenshots or Video _No response_ ### Logs _No response_ ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [√] Flutter (Channel stable, 3.27.2, on Microsoft Windows [version 10.0.19045.5371], locale fr-FR) [√] Windows Version (Installed version of Windows is version 10 or higher) [!] Android toolchain - develop for Android devices (Android SDK version 35.0.0) X cmdline-tools component is missing Run `path/to/sdkmanager --install "cmdline-tools;latest"` See https://developer.android.com/studio/command-line for more details. X Android license status unknown. Run `flutter doctor --android-licenses` to accept the SDK licenses. See https://flutter.dev/to/windows-android-setup for more details. [√] Chrome - develop for the web [X] Visual Studio - develop Windows apps X Visual Studio not installed; this is necessary to develop Windows apps. Download at https://visualstudio.microsoft.com/downloads/. Please install the "Desktop development with C++" workload, including all of its default components [√] Android Studio (version 2023.1) [√] Connected device (4 available) [√] Network resources ! Doctor found issues in 2 categories. ``` </details>
c: crash,platform-android,dependency: dart,will need additional triage
low
Critical
2,806,379,552
angular
Add field to ResourceLoaderParams to find out whether the loader is triggered for the first time
### Which @angular/* package(s) are relevant/related to the feature request? _No response_ ### Description Sometimes, it's necessary to determine whether the loader was triggered for the first time. Before 19.1.x, we could have checked `param.previous.status` for Idle, but now, this status is already set to `loading` initially. ### Proposed solution Hence, I'd like to propose another way to achieve the same, e.g., a `first` flag (similar to the first flag in `ngOnChanges`). ### Alternatives considered We could use a flag in userland. This is doable and can be hidden in a helper function wrapping the loader. However, I assume using a first flag is easier.
area: core,state: has PR,core: reactivity,cross-cutting: signals
low
Minor
2,806,384,481
pytorch
Is there a PyTorch version that can work properly on the Thor platform based on the Blackwell architecture?
Is there a PyTorch version that can work properly on the Thor platform based on the Blackwell architecture? I encounter many errors when compiling PyTorch source code on the Thor platform, and I don't know how to solve them. Is there any expert who can help me cc @ptrblck @msaroufim @eqy
module: cuda,triaged
low
Critical
2,806,403,723
excalidraw
Feature: both a background color and a scratch fill type?
I've noticed that quite often I need to stack shapes on top of one another for some fancier constructions. I use the non-solid fill types to highlight various things, but given that their background is transparent (apart of the shading), stacking them leaks the thing underneath. My current solution is to have 2 of every shape. An "underlay" with a solid white color, and then my actual shape on top with the needed shading. <img width="669" alt="Image" src="https://github.com/user-attachments/assets/08a28964-2eae-4247-a791-f17606115aeb" /> As seen on the above image, this does work, but it needs a lot of effort to duplicate all shapes and handle overlaps between them, so sometimes I end up forgetting or just not wanting to bother (see the outer red shape's 4 corners, each bleeding the black strokes from underneath). Wondering if it would be an interesting idea to add a non-transparent option for the shaded fill type? I'm aware that this would all of a sudden require *2* background colors, so happy to be told no, just thought I'd bring it up as food for thought.
enhancement
low
Minor
2,806,407,068
ui
[bug]: Documentation of toast component for manual installation is wrong
### Describe the bug If you go to the official documentation: https://ui.shadcn.com/docs/components/toast And select manual installation, the three snippets of code that you are supposed to copy and paste into your project are the same one. I guess it was a mistake when creating the page. Could you please fix this and put the correct code for every file? Thanks ### Affected component/components Toast ### How to reproduce If you go to the official documentation: https://ui.shadcn.com/docs/components/toast And select manual installation, the three snippets of code that you are supposed to copy and paste into your project are the same one. I guess it was a mistake when creating the page. ### Codesandbox/StackBlitz link _No response_ ### Logs ```bash ``` ### System Info ```bash All machines, since it is the official website :) ``` ### Before submitting - [x] I've made research efforts and searched the documentation - [x] I've searched for existing issues
bug
low
Critical
2,806,414,251
ollama
deepseek-r1 `qwen` variants use a new pre-tokenizer, which is not implemented in the llama.cpp version used
### What is the issue? The newly supported `deepseek-r1` model variants that have `distill-qwen` in the name use a new pre-tokenizer. Support for this has been added to the latest llama.cpp (not sure if the release version or just the latest commit on the main branch). The backend llama.cpp that Ollama uses should be updated to support this, since the `default` pre-tokenizer is very different than the bespoke version. ### OS Linux, macOS, Windows, Docker, WSL2 ### GPU _No response_ ### CPU _No response_ ### Ollama version 0.5.7
bug
low
Minor
2,806,434,387
pytorch
torch.backends.cudnn.flags use error when test
### 🐛 Describe the bug https://github.com/pytorch/pytorch/blob/main/test/nn/test_convolution.py#L3311 There is a problem with the use of cudnn.flags() here. The original purpose was to test the accuracy of allow_tf32 when it was turned on and off, but the call to cudnn.flags() causes allow_tf32 to always be True. ```python @onlyCUDA @tf32_on_and_off(0.005) def test_ConvTranspose2d_size_1_kernel(self, device): x_cpu = torch.randn(2, 3, 5, 5) conv_cpu = torch.nn.ConvTranspose2d(3, 3, kernel_size=1) y_cpu = conv_cpu(x_cpu) y = torch.rand_like(y_cpu) y_cpu.backward(y) with cudnn.flags(enabled=False): conv_cuda = torch.nn.ConvTranspose2d(3, 3, kernel_size=1).to(device) conv_cuda.bias.data.copy_(conv_cpu.bias.data) conv_cuda.weight.data.copy_(conv_cpu.weight.data) y_cuda = conv_cuda(x_cpu.to(device)) y_cuda.backward(y.to(device)) self.assertEqual(y_cpu, y_cuda, atol=1e-5, rtol=0, exact_device=False) ``` ### Versions Regardless of the environment cc @csarofeen @ptrblck @xwang233 @eqy
module: cudnn,module: convolution,triaged
low
Critical
2,806,442,186
ui
[bug]: Checkbox
### Describe the bug When using the [Tasks demo page](https://ui.shadcn.com/examples/tasks) with Firefox 134.0.2 on Windows 11, the checkbox move when you click on it. See the video below: https://github.com/user-attachments/assets/704cd821-0be9-4322-b2cf-35ce163a2ce7 The problem only reproduces on Firefox. ### Affected component/components Checkbox ### How to reproduce 1. Go to [Tasks demo page](https://ui.shadcn.com/examples/tasks) 2. Click on any checkbox on the first column 3. Observe the checkbox moving up and down ### Codesandbox/StackBlitz link _No response_ ### Logs ```bash ``` ### System Info ```bash Firefox 134.0.2 on Windows 11 ``` ### Before submitting - [x] I've made research efforts and searched the documentation - [x] I've searched for existing issues
bug
low
Critical
2,806,451,493
storybook
[Bug]: can't get scss to load despite storybook does not crash
### Describe the bug I can't get the storybook to show scss styling, I've tried everything (with and without @storybook/preset-scss), but it still doesn't show styling coming from scss files - it does work with inline styling. Storybook doesn't complain. Most webpackFinal config options without @storybook/preset-scss resulted with a storybook error, complaining about scss characters. I have a react app, with webpack5, using scss&css. ``` "@storybook/addon-essentials": "^8.5.0", "@storybook/addon-interactions": "^8.5.0", "@storybook/addon-onboarding": "^8.5.0", "@storybook/addon-webpack5-compiler-swc": "^2.0.0", "@storybook/blocks": "^8.5.0", "@storybook/preset-scss": "^1.0.3", "@storybook/react": "^8.5.0", "@storybook/react-webpack5": "^8.5.0", "style-loader": "4.0.0", "sass": "^1.81.0", "sass-loader": "^16.0.4", "css-loader": "6.11.0", ``` main.ts ``` const config: StorybookConfig = { addons: [ '@storybook/addon-webpack5-compiler-swc', '@storybook/addon-onboarding', '@storybook/addon-essentials', '@chromatic-com/storybook', '@storybook/addon-interactions', '@storybook/preset-scss', ], core: { builder: 'webpack5', }, framework: { name: '@storybook/react-webpack5', options: {}, }, stories: [ '../src/**/*.mdx', '../src/**/*.stories.@(js|jsx|mjs|ts|tsx)', '../src/browser/storybook-utils/**/*.stories.@(js|jsx|mjs|ts|tsx)', ], webpackFinal: async (config) => { config.module.rules.push({ test: /\.scss$/, use: [ 'style-loader', { loader: 'css-loader', options: { importLoaders: 1, // Ensure sass-loader runs before css-loader }, }, 'sass-loader', ], include: path.resolve(path.dirname(import.meta.url), '../'), }); if (config.resolve) { config.resolve.plugins = [ ...(config.resolve.plugins || []), new TsconfigPathsPlugin({ extensions: config.resolve.extensions, }), ]; } return config; }, }; export default config; ``` ``` import './ModalContainerHeader.scss'; ... <div className="ModalContainerHeader"> <div className="testOne" style={{ color: 'red' }}> Test </div> <div className="testTwo">Test</div> </div> ``` ``` .ModalContainerHeader { .testTwo { color: red; } ``` <img width="122" alt="Image" src="https://github.com/user-attachments/assets/36977036-52ed-4086-861b-f7462750fd32" /> <img width="408" alt="Image" src="https://github.com/user-attachments/assets/1f1b3d4c-3a55-49f1-9521-14adb084b481" /> any ideas what to do? ### Reproduction link can't share sorry ### Reproduction steps _No response_ ### System ```bash worked ok ``` ### Additional context _No response_
bug,needs triage
low
Critical
2,806,465,641
neovim
gF should recognize more file/line/column patterns and seek to column
### Problem Right now the [gF](https://neovim.io/doc/user/editing.html#gF) feature seems to only support a limited set of patterns. In particular, it doesn't seem to work with python's backtrace format: ``` Traceback (most recent call last): File "/path/to/file.py", line 123, in <module> ``` It also seems to completely ignore column info since gF on `file:123:12:` (eg as emitted by modern gcc and clang) seeks to the 123rd line but not the 12th column. One of the things I miss most from vscode is that I can run a command in the terminal and just cmd-click in the output when it is referring to a file and go right where I need to go. ### Expected behavior The vscode terminal's logic to detect file/line/column seems to work fairly well in practice, so it may be a good source of inspiration. It looks like [this](https://github.com/microsoft/vscode/blob/ce2c2f3c79a32b9917e32c61e058392dc5a1b6aa/src/vs/workbench/contrib/terminalContrib/links/browser/terminalLinkParsing.ts#L64-L126) is list of 3 regexes that they use, along with examples patterns that they are trying to match. Once you have the parser able to recognize columns, it would be nice if `gF` automatically went to the right column as well. Maybe this should be a new ticket, but it would be really nice if the built-in `:terminal` was automatically enhanced to recognize these patterns and A) apply a highlight group, possibly only when the cursor or mouse is over one (bonus points if the highlight group was only applied when the file exists!), and B) consider a default mapping on `<LeftMouse>` or `<C-LeftMouse>` to go to the file. Another nice feature would be something to send the matching locations to the quickfix list. I have tried `:cgetbuffer` from the terminal and it does not work well, at least with the default `errorformat`. Among other issues, I wasn't able to make one that would match file:line:column in the middle of a line and not treat anything before the file as part of the message rather than the file. It would be especially cool if there was support for only sending the output from the last command to quickfix using OSC codes from the prompt to delimit command boundaries (at least kitty and vscode support the [OSC 133](https://gitlab.freedesktop.org/Per_Bothner/specifications/blob/master/proposals/semantic-prompts.md) A/B/C/D sequence for this which is also useful for #9209). This would let you run your build in a terminal and still get all the niceties like live progress reporting and colorful output, while still being able to use the quickfix list as if you used `:make`
jumps-navigation,editor
low
Critical
2,806,472,718
material-ui
[docs] Tabs need two clicks to work on Firefox
### Related page https://mui.com/material-ui/react-tabs/ ### Kind of issue Broken demo ### Issue description For some examples It needs to click twice to change the tab. That's happening to the firefox browser in chromium based browsers looks fine. ### Context Open the tabs demo documentation in Firefox. I'm on macOS. Going to click to change the tabs, and I'm doing a click on a different tab and looks like clicked, does an effect, stays to the current tab, click again, going to the tab that I clicked. After refresh, it worked, but I cannot consider a cache issue since it was the first time that I visited the page from this browser. **Search keywords**: tabs, double click
docs,component: tabs,status: waiting for author,support: docs-feedback
low
Critical
2,806,540,487
vscode
HTML can't wite any more code Visual Studio Code
Type: <b>Bug</b> I have issue with Visual Studio Code. All of a sudden , i can't write code anymore. let me explain: when i try to mwrite code, it keeps inserting closing tag and i can no longer open any tags. The closing tags appear for all tags like DIV, MAIN, BODY. However ,i've checked my code many times, and it's perfectly fine. I also tried with other project , and i'm experiencing the same problem. I need to solve this issue because i work with VSC. Thank You! VS Code version: Code 1.96.2 (Universal) (fabdb6a30b49f79a7aba0f2ad9df9b399473380f, 2024-12-19T10:22:47.216Z) OS version: Darwin arm64 23.6.0 Modes: <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple M3 (8 x 2400)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off| |Load (avg)|5, 12, 7| |Memory (System)|16.00GB (3.14GB free)| |Process Argv|--crash-reporter-id 8d604037-1790-40a6-96c0-15e093dc2ea9| |Screen Reader|no| |VM|0%| </details>Extensions: none<details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vspor879:30202332 vspor708:30202333 vspor363:30204092 vscod805:30301674 binariesv615:30325510 vsaa593:30376534 py29gd2263:31024239 c4g48928:30535728 azure-dev_surveyone:30548225 962ge761:30959799 pythonnoceb:30805159 pythonmypyd1:30879173 2e7ec940:31000449 pythontbext0:30879054 cppperfnew:31000557 dsvsc020:30976470 pythonait:31006305 dsvsc021:30996838 dvdeprecation:31068756 dwnewjupyter:31046869 newcmakeconfigv2:31071590 nativerepl1:31139838 pythonrstrctxt:31112756 nativeloc1:31192215 cf971741:31144450 iacca1:31171482 notype1:31157159 5fd0e150:31155592 dwcopilot:31170013 stablechunks:31184530 6074i472:31201624 dwoutputs:31217127 hdaa2157:31222309 copilot_t_ci:31222730 ``` </details> <!-- generated by issue reporter -->
info-needed
low
Critical
2,806,557,437
ant-design
AutoComplete customize input size
### Reproduction link [![Edit on CodeSandbox](https://codesandbox.io/static/img/play-codesandbox.svg)](https://codesandbox.io/p/sandbox/customize-input-component-antd-5-23-2-forked-grw9hy?file=%2Fdemo.tsx) ### Steps to reproduce Switch between small, middle and large ### What is expected? AutoComplete wrapper should size to custom Input's height ### What is actually happening? * When size prop is set only on the custom Input, AutoComplete wrapper itself keeps fixed height. * When size prop is set on both Input and AutoComplete,then it works almost as it should, except at large size AutoComplete height is larger the Input's, also it shows the warning in console: Warning: [antd: AutoComplete] You need to control style self instead of setting `size` when using customize input. | Environment | Info | | --- | --- | | antd | 5.23.2 | | React | 18 | | System | Windows | | Browser | Firefox | <!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
improvement
low
Minor
2,806,591,034
storybook
[Bug]: “The current testing environment is not configured to support act” errors when changing between component docs pages
### Describe the bug When switching between component docs pages you get bunch of ``` The current testing environment is not configured to support act(...) ``` and occasionally ``` An update to ZoomElement inside a test was not wrapped in act(...). ``` errors in browser console. ### Reproduction link https://stackblitz.com/edit/github-qczcaoih?file=package.json&preset=node ### Reproduction steps 1. Navigate to Button docs page 2. While on Button docs page, navigate to Header or Page docs page. 3. Check browser console for errors. ### System ```bash System: OS: Linux 5.0 undefined CPU: (8) x64 Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz Shell: 1.0 - /bin/jsh Binaries: Node: 18.20.3 - /usr/local/bin/node Yarn: 1.22.19 - /usr/local/bin/yarn npm: 10.2.3 - /usr/local/bin/npm <----- active pnpm: 8.15.6 - /usr/local/bin/pnpm npmPackages: @storybook/addon-essentials: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/addon-interactions: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/addon-onboarding: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/addon-webpack5-compiler-swc: ^2.0.0 => 2.0.0 @storybook/blocks: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/react: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/react-webpack5: ^8.6.0-alpha.1 => 8.6.0-alpha.1 @storybook/test: ^8.6.0-alpha.1 => 8.6.0-alpha.1 storybook: ^8.6.0-alpha.1 => 8.6.0-alpha.1 ``` ### Additional context React version: 19 Brower: Chrome, Firefox
bug,react,needs triage,upgrade:8.5
low
Critical
2,806,591,543
rust
compiletest: investigate if it's possible to do run-time capability detection against *target*
Some tests are conditioned on *capabilities*, which are most accurately detected at run-time *against* the **target** under cross-compilation scenarios (compared to having to maintain allowlists or denylists of targets). However, implementing this is non-trivial: - You have to build *and* run the capability run-time test as an executable on the **target**, *not* the **host**. This may involve having to work with custom test runners (e.g. `wasmtime`) or remote-test-server. - The *target* may not support std!
E-hard,C-enhancement,A-cross,T-compiler,T-bootstrap,A-compiletest,E-needs-design
low
Minor
2,806,596,749
flutter
[ Web ] Flutter web Autofill doesn't work when browser password manager is protected with Windows Hello
### Steps to reproduce - Create a basic login form with a login and password fields, with autofill - Activate Windows Hello protection on browser (in Chrome : password manager -> settings -> Activate Windows Hello) - First run on browser -> save credentials - Restart your web app - Use this second run to use the saved credentials, with autofill ### Expected results The selected credentials should fill the login/password form fields ### Actual results - [x] if the browser uses its own password manager directly, the autofill works perfectly (tested with Chrome, Mozilla Firefox) - [ ] if the browser uses Windows Hello to protect the password manager, the fields are not filled after Windows validation (tested only with Chrome, because I didnt success to make Mozilla Firefox ask for Windows Hello confirmation) ### Code sample <details open><summary>Code sample</summary> ```dart Form( child: AutofillGroup( child: Column( mainAxisSize: MainAxisSize.min, children: [ //... TextField( autofillHints: const [AutofillHints.username], controller: loginCtrl, ), TextField( autofillHints: const [AutofillHints.password], obscureText: true, controller: pwdCtrl, ), ``` </details>
a: text input,platform-windows,platform-web,has reproducible steps,browser: chrome-desktop,team-web,fyi-text-input,found in release: 3.27,found in release: 3.28
low
Minor
2,806,608,458
kubernetes
Pass ctx to the remaining WaitForXXX methods in test/integration/util/util.go
### What would you like to be added? Pass ctx to the remaining WaitForXXX methods in test/integration/util/util.go ### Why is this needed? We should use existing test context consistently.
sig/scheduling,kind/feature,needs-triage
low
Minor
2,806,630,532
flutter
[A11y]: Slider labels not included in the `Semantics` widget
### Steps to reproduce When creating `Slider` widgets, the `Slider.label` information isn't passed to the semantics widget. ### Expected results Sliders should announce their labels when tabbing through with a screen reader such as NVDA. I can't test platforms other than Windows or Chrome while at work. ### Actual results NVDA announces "Slider", then the percentage. The third slider works perfectly. Looking at the implementation of the `Slider` widget, it seems that there is no `label` property being passed to the `Semantics` widget at the end of the `_buildMaterialSlider` method. I've tested this locally, and adding the `label` argument fixes this. Am I OK to submit a PR with this change? I don't really know how all of this works. I'm keen to help, but - depending on how much vetting I'll need - it might be quicker for someone else to add `label: widget.label` to line 1083 of `slider.dart`. 😄 Thank you for a wonderful framework!! ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; import 'package:flutter/rendering.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); // This widget is the root of your application. @override Widget build(BuildContext context) { final renderBinding = RendererBinding.instance; if (!renderBinding.semanticsEnabled) { renderBinding.ensureSemantics(); } return MaterialApp( title: 'Flutter Demo', theme: ThemeData( colorScheme: ColorScheme.fromSeed(seedColor: Colors.deepPurple), useMaterial3: true, ), home: const MyHomePage(title: 'Flutter Demo Home Page'), ); } } class MyHomePage extends StatefulWidget { const MyHomePage({super.key, required this.title}); final String title; @override State<MyHomePage> createState() => _MyHomePageState(); } class _MyHomePageState extends State<MyHomePage> { int _counter = 0; void _incrementCounter() { setState(() { _counter++; }); } @override Widget build(BuildContext context) { const divider = 100; return Scaffold( appBar: AppBar( backgroundColor: Theme.of(context).colorScheme.inversePrimary, title: Text(widget.title), ), body: Center( child: Column( mainAxisAlignment: MainAxisAlignment.center, children: <Widget>[ const Text( 'You have pushed the button this many times:', ), Text('$_counter'), Slider( autofocus: true, value: _counter / divider, onChanged: (value) => setState( () => _counter = (value * divider).round(), ), label: 'Screen readers will not pick up this label', ), Slider( value: _counter / divider, onChanged: (value) => setState( () => _counter = (value * divider).round(), ), label: 'Screen readers will not speak this label either', semanticFormatterCallback: (value) => 'Screen readers will show this value. Counter is $_counter', ), MergeSemantics( // I use `MergeSemantics` to prevent NVDA from recognising a group. child: Semantics( label: 'This works perfectly', child: Slider( value: _counter / divider, onChanged: (value) => setState( () => _counter = (value * divider).round(), ), label: 'We just need to wrap the `Slider` widget in a `MergeSemantics` and a `Semantics` widget and everything will be fixed', ), ), ) ], ), ), floatingActionButton: FloatingActionButton( onPressed: _incrementCounter, tooltip: 'Increment', child: const Icon(Icons.add), ), ); } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] </details> ### Logs <details open><summary>Logs</summary> ```console ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [Γ£ô] Flutter (Channel master, 3.28.0-2.0.pre.38833, on Microsoft Windows [Version 10.0.26100.2894], locale en-GB) [2.5s] ΓÇó Flutter version 3.28.0-2.0.pre.38833 on channel master at C:\Users\chris\flutter ΓÇó Upstream repository https://github.com/flutter/flutter ΓÇó Framework revision 97ca57cf08 (6 hours ago), 2025-01-23 06:38:20 +0100 ΓÇó Engine revision 97ca57cf08 ΓÇó Dart version 3.8.0 (build 3.8.0-24.0.dev) ΓÇó DevTools version 2.42.0 [Γ£ô] Windows Version (11 Home 64-bit, 24H2, 2009) [4.1s] [Γ£ô] Android toolchain - develop for Android devices (Android SDK version 34.0.0) [4.4s] ΓÇó Android SDK at C:\Users\chris\AppData\Local\Android\sdk ΓÇó Platform android-35, build-tools 34.0.0 ΓÇó Java binary at: C:\Program Files\Android\Android Studio\jbr\bin\java This is the JDK bundled with the latest Android Studio installation on this machine. To manually set the JDK path, use: `flutter config --jdk-dir="path/to/jdk"`. ΓÇó Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-9586694) ΓÇó All Android licenses accepted. [Γ£ô] Chrome - develop for the web [119ms] ΓÇó Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe [Γ£ô] Visual Studio - develop Windows apps (Visual Studio Community 2022 17.12.4) [117ms] ΓÇó Visual Studio at C:\Program Files\Microsoft Visual Studio\2022\Community ΓÇó Visual Studio Community 2022 version 17.12.35707.178 ΓÇó Windows 10 SDK version 10.0.22621.0 [Γ£ô] Android Studio (version 2022.2) [16ms] ΓÇó Android Studio at C:\Program Files\Android\Android Studio ΓÇó Flutter plugin can be installed from: ≡ƒö¿ https://plugins.jetbrains.com/plugin/9212-flutter ΓÇó Dart plugin can be installed from: ≡ƒö¿ https://plugins.jetbrains.com/plugin/6351-dart ΓÇó Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-9586694) [Γ£ô] VS Code, 64-bit edition (version 1.96.2) [14ms] ΓÇó VS Code at C:\Program Files\Microsoft VS Code ΓÇó Flutter extension version 3.102.0 [Γ£ô] Connected device (3 available) [290ms] ΓÇó Windows (desktop) ΓÇó windows ΓÇó windows-x64 ΓÇó Microsoft Windows [Version 10.0.26100.2894] ΓÇó Chrome (web) ΓÇó chrome ΓÇó web-javascript ΓÇó Google Chrome 131.0.6778.265 ΓÇó Edge (web) ΓÇó edge ΓÇó web-javascript ΓÇó Microsoft Edge 132.0.2957.115 [Γ£ô] Network resources [412ms] ΓÇó All expected network resources are available. ΓÇó No issues found! ``` </details>
in triage
low
Major
2,806,638,283
react
Bug: onChange Not Triggering for range input After Dynamic Updates
<!-- Please provide a clear and concise description of what the bug is. Include screenshots if needed. Please test using the latest version of the relevant React packages to make sure your issue has not already been fixed. --> ## Description When using range input in React, the onChange event does not trigger properly under certain dynamic updates to its step and max attributes. Specifically, if the step and max values are updated dynamically, and the range input value is reset to 0, subsequent user interactions with the range slider fail to trigger the onChange event as expected. This behavior does not occur in a pure JavaScript implementation. **React version:** 19.0.0 ## Steps To Reproduce 1. Open the StackBlitz example: [React example](https://stackblitz.com/edit/vitejs-vite-srhuep2z?file=src%2FApp.jsx) 2. Click the middle of the range input, changing its value to `500`. 3. Click the **"Change to step 0.5 max 1"** button, observe that the range input is set to `0` (expected behavior). 4. Click the **"Change to step 500 max 1000"** button, observe that the range input is set back to `0`. Then click the middle of the range input. ## The current behavior After completing step 4, the value of the range input remains at 0, even though the user clicks on the middle of the range slider. The onChange event is not triggered, and the value does not update. ## The expected behavior After completing step 4, the value of the range input should update to 500 when the user clicks on the middle of the range slider. The onChange event should be triggered correctly, reflecting the value change. ## Link to code example - [React implementation reproducing the issue](https://stackblitz.com/edit/vitejs-vite-srhuep2z?file=src%2FApp.jsx) - [Pure JavaScript implementation without the issue](https://codepen.io/wrxue/pen/emOPrbL) --- https://github.com/user-attachments/assets/0d471278-8b12-499d-8b86-819d03a17d2f
Status: Unconfirmed
medium
Critical
2,806,647,618
react
[DevTools Bug]: [DevTools Bug] Cannot reorder children for node "0" because no matching node was found in the Store.
### Website or app https://www.autotrack.nl/aanbod ### Repro steps 1. Start profiling in chrome 2. Click on any filter 3. Uncaught Error: Cannot reorder children for node "0" because no matching node was found in the Store. Dismiss The error was thrown at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1175534 at v.emit (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1140783) at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1142390 at bridgeListener (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1552529) ### How often does this bug happen? Every time ### DevTools package (automated) react-devtools-extensions ### DevTools version (automated) 6.0.1-c7c68ef842 ### Error message (automated) Cannot reorder children for node "0" because no matching node was found in the Store. ### Error call stack (automated) ```text at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1175534 at v.emit (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1140783) at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1142390 at bridgeListener (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1552529) ``` ### Error component stack (automated) ```text ``` ### GitHub query string (automated) ```text https://api.github.com/search/issues?q=Cannot reorder children for node because no matching node was found in the Store. in:title is:issue is:open is:public label:"Component: Developer Tools" repo:facebook/react ```
Type: Bug,Status: Unconfirmed,Component: Developer Tools
medium
Critical
2,806,652,646
rust
pthread_exit crashes on threads created by std::thread::spawn in 1.84, not 1.83, breaking pyo3-log
## Meta Tested on Ubuntu 24.04 and Amazon Linux 2, x86_64. ## Workaround to the production problem In your Cargo.toml that is compiling your Python module, set ```toml [profile.release] debug = 0 lto = false ``` This prevents the 1.84 crashes. However, there is still UB going on even with this setting. ## The Production Problem The crate [pyo3-log](https://docs.rs/pyo3-log/0.12.1/pyo3_log/) installs a bridge that makes [log](https://docs.rs/log/latest/log/trait.Log.html) functions call into Python. This means that all calls to `logging::info!` etc will take the GIL. Python has a stage during interpreter shutdown where attempts to take the GIL will cause a `pthread_exit`. Python 3.14 (still Alpha today, targeted to be released by the end of this year) will change this in python/cpython#87135 - but that will take some time to reach people. This means that if you have a Python program that uses a Rust library and pyo3-log, that spawning a Rust thread, that is calling `logging::info!` in a way unsynchronized with interpreter exit, you'll have unpredicatable crashes in 1.84. ## Minified Program This program: ```rust use std::ffi::c_void; extern "C" { fn pthread_exit(retval: *const c_void); } fn main() { std::thread::spawn(|| { unsafe { pthread_exit(std::ptr::null()); } }); std::thread::sleep(std::time::Duration::from_secs(1)); } ``` when compiled with the following options ``` rustc +1.84 d.rs -Cpanic=abort -Cdebuginfo=limited ``` crashes with this confusing error ``` thread '<unnamed>' panicked at core/src/panicking.rs:223:5: panic in a function that cannot unwind stack backtrace: 0: rust_begin_unwind at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/panicking.rs:665:5 1: core::panicking::panic_nounwind_fmt::runtime at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:119:22 2: core::panicking::panic_nounwind_fmt at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/intrinsics/mod.rs:3535:9 3: core::panicking::panic_nounwind at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:223:5 4: core::panicking::panic_cannot_unwind at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:315:5 5: std::sys::pal::unix::thread::Thread::new::thread_start at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/sys/pal/unix/thread.rs:99:9 6: start_thread 7: clone note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace. thread caused non-unwinding panic. aborting. Aborted ``` This crash happens: 1. Only on 1.84, not on 1.83 2. Only when debuginfo is enabled, but *even if the binary is stripped*. When using `-C panic=unwind` instead, on all versions of the compiler, you get this error: ``` FATAL: exception not rethrown Aborted (core dumped) ``` I seen the claim in Zulip (https://rust-lang.zulipchat.com/#narrow/channel/122651-general/topic/pthread_exit.20from.20a.20Rust-spawned.20thread) that this is undefined behavior, but I'll rather not break pyo3-log
T-compiler,C-bug,T-libs,A-thread,S-has-mcve,S-has-bisection
low
Critical
2,806,660,658
kubernetes
Replace reflect.DeepEqual with cmp.Diff in pkg/scheduler tests
### What would you like to be added? In some tests, `reflect.DeepEqual` is used to compare expected and actual values. It should be replaced to `cmp.Diff` if possible. ### Why is this needed? According to the [SIG Scheduling guidelines](https://github.com/kubernetes/community/blob/master/sig-scheduling/CONTRIBUTING.md#technical-and-style-guidelines), `cmp.Diff` should be used instead of `reflect.DeepEqual`.
sig/scheduling,kind/feature,needs-triage
low
Minor
2,806,663,453
PowerToys
Error after enabling high contrast(accesibility panel) from default unlock window
### Microsoft PowerToys version 0.87.0.0 ### Installation method GitHub ### Running as admin None ### Area(s) with issue? General ### Steps to reproduce lock pc and enable high contrast throw accessibility panel [2025-01-23.txt](https://github.com/user-attachments/files/18519748/2025-01-23.txt) ### ✔️ Expected Behavior No errors on this behavior ### ❌ Actual Behavior Error window displayed ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Critical
2,806,669,948
kubernetes
[Flaking Test] [sig-apps] Kubernetes e2e suite.[It] [sig-apps] ReplicaSet Replace and Patch tests [Conformance]
### Which jobs are flaking? master-blocking - gce-cos-master-default ### Which tests are flaking? Kubernetes e2e suite.[It] [sig-apps] ReplicaSet Replace and Patch tests [Conformance] [Prow](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce/1882170659614756864) [Triage](https://storage.googleapis.com/k8s-triage/index.html?test=ReplicaSet%20Replace%20and%20Patch%20tests%20&xjob=e2e-kops) ### Since when has it been flaking? [1/10/2025, 5:10:10 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-kind-conformance-parallel/1877628702888562688) [1/10/2025, 11:40:02 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce-alpha-enabled-default/1877727102057320448) [1/19/2025, 6:47:20 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/e2e-ci-kubernetes-e2e-al2023-aws-conformance-canary/1880914873404100608) [1/20/2025, 7:25:30 PM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-ubuntu-ec2-containerd/1881467771536019456) [1/22/2025, 5:57:09 PM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce/1882170659614756864) ### Testgrid link https://testgrid.k8s.io/sig-release-master-blocking#gce-cos-master-default ### Reason for failure (if possible) ``` { failed [FAILED] failed to see replicas of test-rs in namespace replicaset-5368 scale to requested amount of 3: watch closed before UntilWithoutRetry timeout In [It] at: k8s.io/kubernetes/test/e2e/apps/replica_set.go:552 @ 01/22/25 21:08:36.337 } ``` ### Anything else we need to know? N/A ### Relevant SIG(s) /sig apps cc: @kubernetes/release-team-release-signal
kind/flake,sig/apps,needs-triage
low
Critical
2,806,699,847
PowerToys
Your message on my computer: A problem has occurred - Report the error in the GitHub repository for PowerToys - 2025-01-23
### Microsoft PowerToys version 0.87.1.0 ### Installation method WinGet ### Running as admin None ### Area(s) with issue? Settings ### Steps to reproduce Date: 23.01.2025 11:34:34 Exception: System.Windows.Markup.XamlParseException: Set property [2025-01-23.txt](https://github.com/user-attachments/files/18519964/2025-01-23.txt) ### ✔️ Expected Behavior _No response_ ### ❌ Actual Behavior ? ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Critical
2,806,712,685
transformers
Paliegemma Pad Token not Masked
### System Info - `transformers` version: 4.48.0 - Platform: Linux-6.1.92-3-x86_64-with-glibc2.31 - Python version: 3.10.15 - Huggingface_hub version: 0.26.2 - Safetensors version: 0.4.5 - Accelerate version: 1.1.1 - Accelerate config: not found - PyTorch version (GPU?): 2.5.1+cu124 (True) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: parallel set-up - Using GPU in script?: Yes - GPU type: NVIDIA A100 80GB PCIe ### Who can help? @amyeroberts @molbap **Description:** I encountered an unexpected behavior while using the Paligemma model. Specifically, the pad tokens' attention masks are being processed to zero (i.e., unmasked status) during the execution of the `self._update_causal_mask(attention_mask, token_type_ids, past_key_values, cache_position, input_ids, inputs_embeds, is_training)` function. After investigating, I found that this behavior stems from a change in the code between the latest version (4.48.0) and the previous version (e.g., 4.43.0). The order in which the pad tokens are masked has been reversed, leading to inconsistent behavior, especially when using left padding. ### Code Comparison **Latest Version (4.48.0):** ```python if attention_mask is not None: causal_mask = causal_mask.clone() # copy to contiguous memory for in-place edit mask_length = attention_mask.shape[-1] padding_mask = causal_mask[:, :, :, :mask_length] + attention_mask[:, None, None, :].to(causal_mask.device) padding_mask = padding_mask == 0 causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( padding_mask, min_dtype ) # we are training thus we need to create a full mask on the image + prefix but causal on suffix if is_training: causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0 ) ``` **Previous Version (4.43.0):** ```python if token_type_ids is not None and labels is not None: # we are training thus we need to create a full mask on the image + prefix but causal on suffix target_length = cache_position[-1] + 1 causal_mask = torch.full( (sequence_length, target_length), fill_value=min_dtype, dtype=dtype, device=device ) if sequence_length != 1: causal_mask = torch.triu(causal_mask, diagonal=1) causal_mask *= torch.arange(target_length, device=device) > cache_position.reshape(-1, 1) causal_mask = causal_mask[None, None, :, :].expand(inputs_embeds.shape[0], 1, -1, -1) if attention_mask is not None: causal_mask = causal_mask.clone() # copy to contiguous memory for in-place edit mask_length = attention_mask.shape[-1] padding_mask = causal_mask[:, :, :, :mask_length] + attention_mask[:, None, None, :].to( causal_mask.device ) # unmask the prefill causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0 ) padding_mask = padding_mask == 0 causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( padding_mask, min_dtype ) ``` **Issue Details** In the latest version (4.48.0), the pad tokens are first masked using the padding_mask: ```python padding_mask = padding_mask == 0 causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( padding_mask, min_dtype ) ``` However, when the following block of code is executed during training: ```python if is_training: causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill( token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0 ) ``` the pad tokens (which have token_type_ids == 0) are unmasked again, effectively reverting the previous masking operation. **Impact** This behavior is particularly problematic when using left padding, as the pad tokens end up being unmasked in the latest version, whereas they were correctly masked in the previous version (4.43.0). **Question** Is this an intentional change, or is it a bug? If it’s a bug, could you please provide guidance on how to address it? I’d be happy to help with a fix if needed. Thanks for your support! ### Information - [x] The official example scripts - [ ] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction ```python from PIL import Image from transformers import AutoProcessor, PaliGemmaForConditionalGeneration processor = AutoProcessor.from_pretrained("path") model = PaliGemmaForConditionalGeneration.from_pretrained("path") prompt = ["<image>caption en","<image>caption en wwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww wwwwwwwwwwwwwwwwwww"] labels = ["123123123123","456456456456"] image_file = ["image path"] * 2 raw_image = [Image.open(i).convert("RGB") for i in image_file] inputs = processor(text=prompt, images=raw_image, suffix=labels, return_tensors="pt", padding="longest") model(**inputs) ``` ### Expected behavior The pad tokens end up being unmasked.
bug,Multimodal,VLM
low
Critical
2,806,743,430
godot
ColorPicker load/save swatches button issues
### Tested versions v4.4.beta1.official [d33da79d3] ### System information Fedora Linux 40 (KDE Plasma) on Wayland - Wayland display driver, Single-window ### Issue description ColorPicker load/save swatches button moves when clicked: https://github.com/user-attachments/assets/e16bfa09-8fb9-492f-acb0-44ae6a18f443 The PopupMenu is also not behaving the same as other Popupmenus in the ColorPicker in RtL: https://github.com/user-attachments/assets/b9a7b0cf-89d0-454a-955f-0d5997194274 There is also an issue in single window mode, where you can't close the menu by clicking on the button again: https://github.com/user-attachments/assets/73ee9d81-e902-4d8a-9c38-975f2da58e21 ### Steps to reproduce Try any ColorPicker in the editor ### Minimal reproduction project (MRP) N/A
bug,topic:gui
low
Minor
2,806,759,283
flutter
revert deprecating Color withOpacity / opacity
Read through the [code](https://github.com/flutter/flutter/blob/master/engine/src/flutter/lib/ui/painting.dart#L328-L332) and [migration guide](https://docs.flutter.dev/release/breaking-changes/wide-gamut-framework) but still having trouble supporting the end result here. ### Axioms 1. the vast majority of production codebases use .withOpacity(x), where x is 0.0–1.0 [1](https://github.com/search?type=code&q=.withOpacity%28) 1. the internal implementation is now .withOpacity(x) => .withAlpha((255.0*x).round()); [2](https://github.com/flutter/flutter/blob/master/engine/src/flutter/lib/ui/painting.dart#L328-L332) 1. the migration docs say to migrate .withOpacity(x) => .withValues(alpha: x); [3](https://docs.flutter.dev/release/breaking-changes/wide-gamut-framework#migrate-withopacity) 1. .withValues(alpha: x) is more verbose and less obvious than .withOpacity(x) [4](https://x.com/spiffyeehaw/status/1882459279564886111?s=46) 1. **designers communicate opacity as a percentage** from 0%–100% not an integer from 0–255 [5](https://help.figma.com/hc/en-us/articles/360041098433-Adjust-the-properties-of-an-image#:~:text=Click%20the%20thumbnail%20to%20open,by%20clicking%20the%20eye%20icon.) ### Questions 1. why doesn't the internal implementation match the migration docs 1. why does withAlpha(x) take an int 0–255 and withValues(alpha: x) take a double 0–1.0 1. why does .withOpacity need to be deprecated at all ### Recommendations 1. unify alpha to a double or an int across all color methods 1. **undeprecate withOpacity** method 1. **undeprecate opacity** getter
engine,customer: crowd,c: proposal,P3,team-engine,triaged-engine
high
Critical
2,806,782,186
material-ui
[Chip] Clickable area expands with `line-height`
### Steps to reproduce Steps: 1. Open this link to live example: https://stackblitz.com/edit/react-ryuspuxm?file=Demo.tsx 2. Click anywhere inside the black border box ### Current behavior The chip is clicked even though the pointer is outside the Chip. ### Expected behavior The chip should only be clicked when the pointer is inside the Chip. ### Context Found in https://github.com/mui/mui-x/issues/16304 ### Your environment <details> <summary><code>npx @mui/envinfo</code></summary> ``` System: OS: macOS 14.5 Binaries: Node: 22.2.0 - ~/.nvm/versions/node/v22.2.0/bin/node npm: 10.7.0 - ~/.nvm/versions/node/v22.2.0/bin/npm pnpm: 9.15.4 - ~/.nvm/versions/node/v22.2.0/bin/pnpm Browsers: Chrome: 131.0.6778.265 Edge: 132.0.2957.115 Safari: 17.5 npmPackages: @mui/material: 6.4.1 ``` </details> **Search keywords**: Chip line-height
bug 🐛,component: chip,ready to take
low
Minor
2,806,782,948
flutter
iOS Zoom is very sensitive
### Steps to reproduce 1. Open the camera on an iOS device 2. Pinch zoom ### Expected results Zooming in using the pinch gesture functions the same as the native camera. ### Actual results Zooming in using the pinch gesture is extremely sensitive & zooms in a lot, very quickly. ### Code sample <details open><summary>Code sample</summary> ```dart class CameraZoom { final double minimunCameraZoom; final double maximumCameraZoom; final double currentCameraZoom; final double zoomScale; CameraZoom({ required this.minimunCameraZoom, required this.maximumCameraZoom, required this.currentCameraZoom, required this.zoomScale, }); bool isZoomOut() { return zoomScale < 1.0; } double calculateZoomOut() { final zoomToApply = zoomScale < 1.0 ? currentCameraZoom - (maximumCameraZoom / 200) : currentCameraZoom; return zoomToApply; } double calculateZoomIn() { final zoomToApply = zoomScale > 1.0 ? currentCameraZoom + (maximumCameraZoom / 200) : currentCameraZoom; return zoomToApply; } double calculateZoomToApply() { double scale = isZoomOut() ? calculateZoomOut() : calculateZoomIn(); if (scale <= minimunCameraZoom) { scale = minimunCameraZoom; } else if (scale >= maximumCameraZoom) { scale = maximumCameraZoom; } return scale; } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> https://github.com/user-attachments/assets/4c0f94de-2daf-45be-9b89-95abf9f86b36 </details> ### Logs <details open><summary>Logs</summary> ```console [Paste your logs here] ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [Paste your output here] ``` </details>
waiting for customer response,in triage
low
Minor
2,806,796,663
flutter
[iOS][PlatformView] Conflict between Flutter ListView scrolling and PlatformView
### Steps to reproduce This example highlights a conflict between Flutter's scrolling and the native platform view's `TextView` scrolling and text selection. Specifically, text selection in the TextView only works when `ListView` scrolling is disabled using physics: NeverScrollableScrollPhysics(). This limitation affects the usability of TextView and could also impact other types of platform views. **Note:** This issue occurs only on iOS. ### Expected results Users should be able to select text even when the TextView is inside a scrollable widget. ### Actual results User cannot select text when TextView is inside a scrollable Widget and scrolling is enabled. ### Code sample [flutter_native_text_view_sample_main.zip](https://github.com/user-attachments/files/18545585/flutter_native_text_view_sample_main.zip) <details open><summary>Code sample</summary> **Run this ready code:** Screen has 3 Texts elements: - the first one is flutter widget - the other two are platform views. Try to select text, it works with flutter widget but doesn't work with platform views. If scrolling is disable, text selection works also with platform views. After testing we think there is a conflict at low level. Sample here: </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> - Flutter's SelectableText: Text selection works as expected. - Native Platform View: Text selection is hindered by scrolling functionality. https://github.com/user-attachments/assets/f7abf303-97b7-4d74-89cc-5ce17426dcae </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```Doctor summary (to see all details, run flutter doctor -v): [!] Flutter (Channel [user-branch], 3.24.5, on macOS 14.2.1 23C71 darwin-arm64, locale en-IT) ! Flutter version 3.24.5 on channel [user-branch] at /Users/delfme/Development/flutter Currently on an unknown channel. Run `flutter channel` to switch to an official channel. If that doesn't fix the issue, reinstall Flutter by following instructions at https://flutter.dev/setup. ! Upstream repository unknown source is not a standard remote. Set environment variable "FLUTTER_GIT_URL" to unknown source to dismiss this error. [✓] Android toolchain - develop for Android devices (Android SDK version 34.0.0) [✓] Xcode - develop for iOS and macOS (Xcode 15.2) [✓] Chrome - develop for the web [✓] Android Studio (version 2024.1) [✓] Connected device (4 available) ! Error: Browsing on the local area network for iPhone. Ensure the device is unlocked and attached with a cable or associated with the same local area network as this Mac. The device must be opted into Developer Mode to connect wirelessly. (code -27) [✓] Network resources ``` </details>
platform-ios,framework,f: scrolling,a: platform-views,has reproducible steps,team-ios,found in release: 3.27,found in release: 3.29
low
Critical
2,806,796,688
node
[v20.x] `parallel/test-buffer-tostring-range` fails on SmartOS
### Test parallel/test-buffer-tostring-range ### Platform SmartOS ### Console output ```console 03:13:42 not ok 177 parallel/test-buffer-tostring-range 03:13:42 --- 03:13:43 duration_ms: 5550.52900 03:13:43 severity: fail 03:13:43 exitcode: 1 03:13:43 stack: |- 03:13:43 node:internal/buffer:961 03:13:43 super(bufferOrLength, byteOffset, length); 03:13:43 ^ 03:13:43 03:13:43 RangeError: Array buffer allocation failed 03:13:43 at new ArrayBuffer (<anonymous>) 03:13:43 at new Uint8Array (<anonymous>) 03:13:43 at new FastBuffer (node:internal/buffer:961:5) 03:13:43 at Function.alloc (node:buffer:395:10) 03:13:43 at Object.<anonymous> (/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos23-x64/test/parallel/test-buffer-tostring-range.js:107:28) 03:13:43 at Module._compile (node:internal/modules/cjs/loader:1468:14) 03:13:43 at Module._extensions..js (node:internal/modules/cjs/loader:1547:10) 03:13:43 at Module.load (node:internal/modules/cjs/loader:1287:32) 03:13:43 at Module._load (node:internal/modules/cjs/loader:1103:12) 03:13:43 at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:180:12) 03:13:43 03:13:43 Node.js v20.18.3 03:13:43 ... ``` ### Build links - https://ci.nodejs.org/job/node-test-commit-smartos/58746/console ### Additional information It's blocking v20 release https://github.com/nodejs/node/pull/56699 @nodejs/platform-smartos
smartos,flaky-test,v20.x
low
Critical
2,806,815,184
vscode
Braces colouring is off inside Dart multiline strings
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.96.4 - OS Version: Windows_NT x64 10.0.22631 Steps to Reproduce: 1. Open an empty folder with nothing inside 2. Disable all extensions 3. Add a new `file.dart` file 4. Write the following to it: ```dart var myString = ''' Here is a string with () and some other examples {like (this)}. '''; ``` 5. See the braces colouring (possibly, more on that later) ![Image](https://github.com/user-attachments/assets/87fbf68b-4dbb-43b0-bbba-f070e35ceab5) 6. Change the string to be raw by adding `r` before the opening `'''`: ![Image](https://github.com/user-attachments/assets/b70b8bbc-ad07-43a8-8eee-a2216a434a4f) I've asked @DanTup to verify this but with the same VS Code version he could not see this behaviour on his machine. If I close the editor (file opened) and reopen it (even removing `r` now) I see the braces colouring flash and go back to string colour. His guess is that there may be some other racing here as https://github.com/microsoft/vscode/issues/230498. Same machine, a Samsung Galaxy Book Pro. To note we had problems with this colouring previously at https://github.com/microsoft/vscode/issues/209435 for Dart and we still have this opened for Python https://github.com/microsoft/vscode/issues/225179. I can't reliably reproduce this (besides flashing) again on this _exact_ folder/file (unfortunatelly). But I saw this behaviour previously on Dart projects and this is an empty folder with only that file inside.
triage-needed
low
Critical
2,806,817,554
react-native
SectionList onViewableItemsChanged behaviour difference between iOS and Android/Web
### Description **Expected:** `SectionList` `onViewableItemsChanged` should correctly report `viewableItems` across multiple sections when items from those sections are within the viewport. This is the current behaviour on Android and Web. **Actual:** `SectionList` `onViewableItemsChanged` will only report `viewableItems` for one section at at time, even if items from multiple sections are within the viewport. This is the current behaviour on iOS. Notes: * Using `getItemLayout` restores the expected behaviour on iOS. * The reproducer uses the `viewableItems` parameter. However, the issue is also true of the `changed` parameter. * I've tested the scenarios with and without passing varying `viewabilityConfig`. * Not the same behaviour as reported in https://github.com/facebook/react-native/issues/48351. The following statement from that issue is not true of this issue: "If I remove getItemLayout entirely, the logic works fine regardless." ### Steps to reproduce 1. Open the supplied Expo reproducer 2. Scroll through the list 3. Observe the logs (messages list the sections with viewable items) 4. Do the above across multiple platforms, when scrolled to a point where multiple sections are within the viewport: 1. iOS will incorrectly only report a single section. 2. Android and Web will correctly report multiple sections. Reproduced locally in vanilla `[email protected]` with new architecture. Also reproduced locally in `[email protected]` without new architecture. ### React Native Version 0.76.3 ### Affected Platforms Runtime - iOS ### Output of `npx react-native info` ```text System: OS: macOS 15.2 CPU: (12) arm64 Apple M2 Pro Memory: 139.78 MB / 32.00 GB Shell: version: "5.9" path: /bin/zsh Binaries: Node: version: 20.12.2 path: ~/.nvm/versions/node/v20.12.2/bin/node Yarn: version: 1.22.19 path: /opt/homebrew/bin/yarn npm: version: 10.5.0 path: ~/.nvm/versions/node/v20.12.2/bin/npm Watchman: Not Found Managers: CocoaPods: version: 1.16.2 path: /Users/philipbulley/.rbenv/shims/pod SDKs: iOS SDK: Platforms: - DriverKit 24.2 - iOS 18.2 - macOS 15.2 - tvOS 18.2 - visionOS 2.2 - watchOS 11.2 Android SDK: Not Found IDEs: Android Studio: 2024.2 AI-242.23339.11.2421.12700392 Xcode: version: 16.2/16C5032a path: /usr/bin/xcodebuild Languages: Java: version: 17.0.10 path: /usr/bin/javac Ruby: version: 3.3.0 path: /Users/philipbulley/.rbenv/shims/ruby npmPackages: "@react-native-community/cli": installed: 15.0.1 wanted: 15.0.1 react: installed: 18.3.1 wanted: 18.3.1 react-native: installed: 0.76.3 wanted: 0.76.3 react-native-macos: Not Found npmGlobalPackages: "*react-native*": Not Found Android: hermesEnabled: true newArchEnabled: true iOS: hermesEnabled: true newArchEnabled: true ``` ### Stacktrace or Logs ```text n/a ``` ### Reproducer https://snack.expo.dev/@pbcodes/onviewableitemschanged-behaviour-difference-between-ios-and-android-or-web ### Screenshots and Videos ### iOS ❌ Scroll until items from 2 sections are within viewport <img width="1366" alt="Image" src="https://github.com/user-attachments/assets/117fdcbf-b4bf-4003-b560-7659b9d8d177" /> Scroll further until the first section has left the viewport <img width="1366" alt="Image" src="https://github.com/user-attachments/assets/a3a970f3-9198-4adc-9f8c-f47dc5401509" /> ### Android ✅ Scroll until items from 2 sections are within viewport <img width="1366" alt="Image" src="https://github.com/user-attachments/assets/e2bf13ec-2069-4de5-81e6-a8a499b34ba3" /> ### Web ✅ Scroll until items from 2 sections are within viewport <img width="1366" alt="Image" src="https://github.com/user-attachments/assets/043bc864-90d5-4226-b242-01a919c7fa33" />
Platform: iOS,Platform: Android,Component: SectionList
low
Minor
2,806,829,033
pytorch
[custom ops] [2.7 nightly] custom ops with typing.List breaks when importing annotations from future
### 🐛 Describe the bug The latest nightly version does not support `typing.List` in combination with `from __future__ import annotations`, but requires the use of `list`. While I agree that using `list` is better and more modern, this introduces a breaking change that makes it difficult to keep the custom ops working both on older and newer pytorch versions. This change was introduced 5 days ago with this commit: https://github.com/pytorch/pytorch/commit/a79100ab11e09473d7456bc02de36de629d9db62#diff-09dd038b7467c407d31c7feb905906037d68d4a419a714bb1f56cee93976de39 My problem with that is: Using List[T] as annotation now gives me the following error: ``` Traceback (most recent call last): File "/weka/mathias/kuma/.venv/lib/python3.12/site-packages/torch/_library/infer_schema.py", line 65, in convert_type_string return eval(annotation_type) ^^^^^^^^^^^^^^^^^^^^^ File "<string>", line 1, in <module> NameError: name 'List' is not defined. Did you mean: 'list'? ``` I could simply change all annotations for `list`, but then it doesn't work anymore with older pytorch versions, requring to create a complete clone of the custom op module. I already tried using typing Aliases: use `type List = list if version... else type List = typing.List`, but that doesn't work either, because the types are evaluated as a raw string - using capital L in the beginning doesn't work. **Turns out that typing.List only fails when importing future annotations.** So, not sure if you want to fix, but since it did work before and would only require reverting one line of code, I wanted to open this issue to discuss if you want to support this case or not. My workaround now was to remove the `from __future__ import annotations` from these files and change some types. The line that need to be changed https://github.com/pytorch/pytorch/blob/a79100ab11e09473d7456bc02de36de629d9db62/torch/_library/infer_schema.py#L6 to the state before this commit: https://github.com/pytorch/pytorch/commit/a79100ab11e09473d7456bc02de36de629d9db62#diff-09dd038b7467c407d31c7feb905906037d68d4a419a714bb1f56cee93976de39 Use this script to reproduce the error: ``` from __future__ import annotations import torch from typing import List @torch.library.custom_op("test::custom_op_list", mutates_args=()) def my_custom_op( x: torch.Tensor, ) -> List[torch.Tensor]: return [torch.randn_like(x)] ``` ### Versions Pytorch 2.7.0.dev20250122+cu124 Python 3.12.6 Ubuntu 20.04 (collect_env.py crashes when using uv) cc @ezyang @gchanan @zou3519 @kadeng @msaroufim @chauhang @penguinwu @bdhirsh @yf225
high priority,triaged,module: custom-operators,oncall: pt2,module: pt2-dispatcher
low
Critical
2,806,850,843
vscode
We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. We were unable to automatically build your code. Please change the build mode for this language to manual and specify build steps for your project. See https://docs.github.com/en/code-security/code-scanning/troubleshooting-code-scanning/automatic-build-failed for more information. Encountered a fatal error while running "/opt/hostedtoolcache/CodeQL/2.20.1/x64/codeql/codeql database trace-command --use-build-mode --working-dir /home/runner/work/todo/todo -O=cpp.trap.cache.dir=/home/runner/work/_temp/trapCaches/cpp -O=cpp.trap.cache.bound=1024 -O=cpp.trap.cache.write=true /home/runner/work/_temp/codeql_databases/cpp". Exit code was 2 and error was: A fatal error occurred: Exit status 1 from command: [/opt/hostedtoolcache/CodeQL/2.20.1/x64/codeql/cpp/tools/autobuild.sh]. See the logs for more details.
https://github.com/latamexiovas10/todo/actions/runs/12929318719/job/36058459946#step:7:161
info-needed,triage-needed
low
Critical
2,806,857,981
react
CSP violation
As per guidelines from security team, framework team wants to enable Content Security Policy (CSP ) at the gateway layer to enforce good web-security practice. In 2412, framework team evaluated the impact of enabling lenient CSP at the gateway layer. As part of this exercise, we enabled CSP in the gateway layer in Report only mode and collected all CSP violations. we found that node_modules/react-dom/cjs/react-dom-server-legacy.browser.development.js node_modules/react-dom/cjs/react-dom-server-legacy.browser.production.min.js node_modules/react-dom/cjs/react-dom-server-legacy.node.development.js node_modules/react-dom/cjs/react-dom-server.browser.development.js node_modules/react-dom/cjs/react-dom-server.node.development.js node_modules/react-dom/umd/react-dom-server-legacy.browser.development.js node_modules/react-dom/umd/react-dom-server-legacy.browser.production.min.js node_modules/react-dom/umd/react-dom-server.browser.development.js is using eval/dynamic function which is flagged as CSP violation. To fix this, we need to replace usage of eval/dynamic function with something safe.
Status: Unconfirmed
medium
Minor
2,806,867,847
vscode
Allow multiple lines per StickyScroll entry
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Some coding conventions mandate opening/closing braces to be on a separate line. For example here's a piece of `PostgreSQL` code: <details> <summary>Wall of C code</summary> ```C RelFileNumber GetNewRelFileNumber(Oid reltablespace, Relation pg_class, char relpersistence) { RelFileLocatorBackend rlocator; char *rpath; bool collides; ProcNumber procNumber; /* * If we ever get here during pg_upgrade, there's something wrong; all * relfilenumber assignments during a binary-upgrade run should be * determined by commands in the dump script. */ Assert(!IsBinaryUpgrade); switch (relpersistence) { case RELPERSISTENCE_TEMP: procNumber = ProcNumberForTempRelations(); break; case RELPERSISTENCE_UNLOGGED: case RELPERSISTENCE_PERMANENT: procNumber = INVALID_PROC_NUMBER; break; default: elog(ERROR, "invalid relpersistence: %c", relpersistence); return InvalidRelFileNumber; /* placate compiler */ } .... } ``` </details> The StickyScroll makes reading this kind of code really nice, but sometimes there are extremely long `if` statements and having those Stick is really nice. However the `outlineModel` doesn't really care about the `if` statements. The next best alternative would be the `foldingProviderModel`, but that doesn't work well with this coding convention. The Sticky looks like this: ```C { { { /* some code here */ ... ``` And surely if one were to fold the opening braces to the same line, it'd look way more useful: ```C GetNewRelFileNumber(Oid reltablespace, Relation pg_class, char relpersistence){ switch (relpersistence){ /* you are here */ ``` ## Feature: Allow including "lines before" and "lines after" into the sticky scroll, or a hack to "collapse braces" or something, although that is too language specific, unfortunately.
feature-request,editor-sticky-scroll
low
Critical
2,806,891,699
react
[React 19] Without extra Generic Type to useActionState hook
## Summary Hi everyone! I have a question about the types from the new hook 'useActionState'. Actually, I'm working on a project and I need some extra types for my action function ![A snippet of types from React 19v](https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bxewm33ed1v9kqmgceq8.png) Why don't we have an extra Generic Type to return from the action? Does the return really need to be equal to the initial state? In a real use case, the return from the action can be different from the entry data, such as an error object like: ```js // INPUT DATA { email: '[email protected]' } // OUTPUT DATA { email: { value: '[email protected]' state: { error: { message: null } } } } ``` **Reference** - https://react.dev/reference/react/useActionState - https://react.dev/blog/2024/12/05/react-19#whats-new-in-react-19 <!-- Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a repository on GitHub, or provide a minimal code example that reproduces the problem. You may provide a screenshot of the application if you think it is relevant to your bug report. Here are some tips for providing a minimal example: https://stackoverflow.com/help/mcve. -->
React 19
medium
Critical
2,806,902,523
PowerToys
Last Modifed Date Is Sometimes More Appropriate Than File Creation Date
### Description of the new feature / enhancement I shoot many videos with my camera. Then I remove the memory card, insert into a reader on my Windows computer, and use the Sony utility to copy them to a Windows volume. The copy modifies the file creation date. If I let videos accumulate for a few days before copying them to Windows, they all end up with the same creation date. However, the last modified date remains correct, which is the date each video was made. ### Scenario when this would be used? It might be appropriate to use file creation date in some circumstances, but anyone who is copying media from a memory card to their computer will want to use last modified date instead. This should be an option. ### Supporting information * https://www.mslinn.com/git/1600-git-ls-files.html#ucff * https://support.d-imaging.sony.co.jp/www/disoft/int/download/playmemories-home/win/en/index.html
Needs-Triage
low
Minor
2,806,914,239
next.js
pnpm workspace package with non shared lockfile break dev mode
### Link to the code that reproduces this issue https://github.com/LukasGerm/pnpm-next-test ### To Reproduce 1. pnpm i 2. Go into the packages/test-package and run pnpm build 3. Go to apps/my-app and run pnpm dev ### Current vs. Expected behavior Observe that <img width="397" alt="Image" src="https://github.com/user-attachments/assets/e4808a47-5c0e-4e65-b1e1-e26bf882878f" /> Expected would be that this works. Building the application works btw. ### Provide environment information ```bash Operating System: Platform: darwin Arch: arm64 Version: Darwin Kernel Version 24.1.0: Thu Oct 10 21:06:57 PDT 2024; root:xnu-11215.41.3~3/RELEASE_ARM64_T6041 Available memory (MB): 24576 Available CPU cores: 12 Binaries: Node: 20.9.0 npm: 10.1.0 Yarn: N/A pnpm: 9.15.4 Relevant Packages: next: 15.1.6 // Latest available version is detected (15.1.6). eslint-config-next: 15.1.6 react: 19.0.0 react-dom: 19.0.0 typescript: 5.7.3 Next.js Config: output: N/A ``` ### Which area(s) are affected? (Select all that apply) Turbopack ### Which stage(s) are affected? (Select all that apply) next dev (local) ### Additional context I also tested without turbopack, but the same issue
Turbopack,linear: turbopack,Package Managers
low
Minor
2,806,923,162
flutter
FlutterEngineGroup isolates run in same thread but should not
### Steps to reproduce I am utilizing FlutterEngineGroup between my main and my worker isolates on mobile devices so that I can share [complex objects](https://github.com/flutter/flutter/issues/152484). I am also using a fork of flutter_isolate plugin, so that I can have callbacks from native side to the respective isolate. With their recent change, it looks like all the isolates which are spawned from the same FlutterEngineGroup, run in the same thread. I verified this by calling `syscall(SYS_gettid)` via FFI and it came back with same numbers. More context [here](https://github.com/rmawatson/flutter_isolate/issues/160). Is this expected? Do I need to spawn a separate native thread in order to run the new engine (within same engine group) in a separate thread? ### Expected results Isolates within same FlutterEngineGroup should run in separate threads ### Actual results Isolates within same FlutterEngineGroup run in same thread. ### Code sample <details open><summary>Code sample for retrieving thread id</summary> ```c++ #include <unistd.h> #include <sys/syscall.h> #include <stdint.h> int64_t get_thread_id() { return (int64_t)syscall(SYS_gettid); } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] </details> ### Logs <details open><summary>Logs</summary> ```console [Paste your logs here] ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console [Paste your output here] ``` </details>
waiting for customer response,in triage
low
Major
2,806,939,325
flutter
[file_selector] Add option to customize the directory selection dialog title
### Use case When calling `getDirectoryPath` I couldn't find a way to customize the dialog title. This would be important to give context to the user about the purpose of the directory selection. ### Proposal A ppssible solution is to add an optional parameter to the method: `getDirectoryPath({String? dialogTitle})`
package,team-ecosystem,P3,p: file_selector,triaged-ecosystem
low
Minor
2,806,954,203
godot
Unrecognized UID: "uid://a"
### Tested versions 4.4 beta1 ### System information W10 ### Issue description When opening project ``` Unrecognized UID: "uid://a" ``` is printed to the console. I tracked the source to GDScript parser, but no idea what causes it. ### Steps to reproduce Open MRP. ### Minimal reproduction project (MRP) [Godot-Instance-Dock.zip](https://github.com/user-attachments/files/18521292/Godot-Instance-Dock.zip) Includes .godot folder, because the error is rather finicky. There is a chance you still won't be able to reproduce it.
bug,topic:gdscript,needs testing,regression
low
Critical
2,806,971,341
ui
[bug]: When trying to initialize `shadcn-ui` in an Astro project, I encounter an error: `ENOENT: no such file or directory`, related to the creation of the folder `src/*/lib`.
### Describe the bug ### Description When trying to initialize `shadcn-ui` in an Astro project, I encounter an error: `ENOENT: no such file or directory`, related to the creation of the folder `src/*/lib`. > ENOENT: no such file or directory, mkdir 'C:_beppe\Astro_dev\ahorros-familiares-astro-sphere\src*\lib' of my project I already have a project created based on the AstroTheme called 'Astro-sphere'. I'm not sure if this could be causing the issue when installing shadcn on top of it. I have tried creating the folders manually, but it still doesn't work. What surprises me is that it mentions a path like src/*/lib, but I don't have this structure in my Astro project. I'm not sure what the * is supposed to refer to ### Affected component/components Installation on Astro ### How to reproduce ### Steps to reproduce 1. Create an Astro project using `npm create astro@latest`. 2. Configure Tailwind CSS according to the documentation. 3. Run the command `npx shadcn@latest init`. 4. Choose the default options and force dependency installation (`--force`). 5. The following error occurs: ### Codesandbox/StackBlitz link _No response_ ### Logs ```bash PS C:\_beppe\Astro\_dev\ahorros-familiares-astro-sphere> npx shadcn@latest init ✔ Preflight checks. ✔ Verifying framework. Found Astro. ✔ Validating Tailwind CSS. ✔ Validating import alias. √ Which style would you like to use? » New York √ Which color would you like to use as the base color? » Neutral √ Would you like to use CSS variables for theming? ... no / yes ✔ Writing components.json. ✔ Checking registry. ✔ Updating tailwind.config.mjs ✔ Updating src\styles\global.css Installing dependencies. It looks like you are using React 19. Some packages may fail to install due to peer dependency issues in npm (see https://ui.shadcn.com/react-19). √ How would you like to proceed? » Use --force ✔ Installing dependencies. ⠹ Updating files. Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. ENOENT: no such file or directory, mkdir 'C:\_beppe\Astro\_dev\ahorros-familiares-astro-sphere\src\*\lib' ``` ### System Info ```bash ### Additional details - **Operating system**:Windows 10 - **Node.js version**: v20.11.0 - **shadcn-ui version**: Latest (`npx shadcn@latest`) - **Astro version**: "astro": "^4.15.12", - **Tailwind CSS version**: "^3.4.13" ``` ### Before submitting - [x] I've made research efforts and searched the documentation - [x] I've searched for existing issues
bug
low
Critical
2,806,976,716
electron
document.fragmentDirective has no properties inside webview
### Preflight Checklist - [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success. ### Electron Version 34.0.1 ### What operating system(s) are you using? Ubuntu ### Operating System Version 23.04 ### What arch are you using? x64 ### Last Known Working Electron version _No response_ ### Expected Behavior When calling `document.fragmentDirective`, it should have an `items` and a `createSelectorDirective` property[1]. This is the case when calling it from a browserWindow's console or webcontentsview, [1] details https://github.com/WICG/scroll-to-text-fragment/blob/main/fragment-directive-api.md ### Actual Behavior When calling it from a webview's console then `document.fragmentDirective` has no properties, which mean you can't create text fragment links from pages inside webviews ### Testcase Gist URL _No response_ ### Additional Information This is a relatively new API.
component/webview,bug :beetle:,platform/all,has-repro-gist,34-x-y,35-x-y
low
Critical
2,806,989,174
terminal
User32.dll Import Cause Windows Terminal Crash
### Windows Terminal version 1.21.3231.0 ### Windows build number 10.0.26100.0 ### Other Software _No response_ ### Steps to reproduce ## Problem Statement It seems like the use of user32.dll could cause stability issue on Windows terminal. Eventually, it will cause the crash. This issue only happens with a small possibility when the command is executed (<0.1% chance/cycle). But when we put this in a while loop a run the same script across 200+ machines, it will happen everyday. In this case, I isolated the code to reproduce this problem. * run following code as administrator. * start another 15 terminals to run that in parallel. ## Code to reproduce ```powershell function Out-ProcessEx { param ( [Parameter(Mandatory, ValueFromPipeline)] $process ) process { #Get Window Title, We use this function because default PowerShell function cannot get #title for hidden window $wTitle = Get-WindowTitleFromProcessId -processId $process.Id # update MainWindowTitle $field = $process.GetType().GetField("mainWindowTitle", [System.Reflection.BindingFlags]::Instance -bor [System.Reflection.BindingFlags]::NonPublic -bor [System.Reflection.BindingFlags]::Public) $field.SetValue($process, $wTitle) $process } } function Get-WindowHandleFromProcessId { param ( [Parameter(Mandatory = $true)] [int]$ProcessId ) # Define the necessary functions from user32.dll Add-Type @" using System; using System.Runtime.InteropServices; public class User32 { public delegate bool EnumWindowsProc(IntPtr hWnd, IntPtr lParam); [DllImport("user32.dll", SetLastError = true)] [return: MarshalAs(UnmanagedType.Bool)] public static extern bool EnumWindows(EnumWindowsProc lpEnumFunc, IntPtr lParam); [DllImport("user32.dll", SetLastError = true)] public static extern int GetWindowThreadProcessId(IntPtr hWnd, out int lpdwProcessId); } "@ -ErrorAction SilentlyContinue $Script:windowHandle = 0 $callback = [User32+EnumWindowsProc] { param ($hWnd, $lParam) $processId = 0 [User32]::GetWindowThreadProcessId($hWnd, [ref]$processId) | Out-Null if ([int]($processId) -eq [int]($lParam)) { $Script:windowHandle = $hWnd return $false # Stop enumeration } return $true # Continue enumeration } [User32]::EnumWindows($callback, [IntPtr]$ProcessId) | Out-Null return $Script:windowHandle } function Get-WindowTitleFromProcessId { param ( [Parameter(Mandatory = $true)] [Int]$processId ) # Define the necessary functions from user32.dll Add-Type @" using System; using System.Text; using System.Runtime.InteropServices; public class WindowTitleFromHandle { [DllImport("user32.dll", CharSet = CharSet.Auto, SetLastError = true)] public static extern int GetWindowText(IntPtr hWnd, StringBuilder lpString, int nMaxCount); [DllImport("user32.dll", SetLastError = true)] public static extern int GetWindowTextLength(IntPtr hWnd); } "@ -ErrorAction SilentlyContinue $WindowHandle = Get-WindowHandleFromProcessId -ProcessId $processId $length = [WindowTitleFromHandle]::GetWindowTextLength($WindowHandle) if ($length -gt 0) { $builder = New-Object System.Text.StringBuilder -ArgumentList ($length + 1) [WindowTitleFromHandle]::GetWindowText($WindowHandle, $builder, $builder.Capacity) | Out-Null return $builder.ToString() } else { return "No title found" } } while($true) { $s1 = Get-Process|Out-ProcessEx #start 16 process and see...... } ``` ## Error Log ![Image](https://github.com/user-attachments/assets/e98509ef-bf6a-44c6-acc3-54ea18276cb7) ![Image](https://github.com/user-attachments/assets/7ed03a72-3995-4db8-b1d1-5a824d4ef95b) ### Expected Behavior 16 instances of windows terminal should persist after 20 minutes ### Actual Behavior * The crash issue typically occur after ~ 15-20mins.
Issue-Bug,Needs-Triage
low
Critical
2,806,995,030
terminal
Can't disable startup
### Windows Terminal version 1.21.3231.0 ### Windows build number 10.0.26100.0 ### Other Software Docker (disabled on startup, does not launch) AthenaOS WSL ### Steps to reproduce 1. Install WSL (any distro, currently using AthenaOS WSL) 2. Install Windows Terminal 3. Install Docker Desktop 4. Disable docker on startup 5. Disable Launch on Machine startup in Windows Terminal settings ![Image](https://github.com/user-attachments/assets/ad149f3c-6184-403a-b7c2-e0418a81931f) 7. Check `shell:startup` ![Image](https://github.com/user-attachments/assets/5a08cd1a-c5d1-4d01-a241-90a89295b5c9) 8. Check `shell:Common Startup` ![Image](https://github.com/user-attachments/assets/1dd7ffb1-0aca-4dec-ad5f-cb72e9a63110) 9. Check task manager startup apps ![Image](https://github.com/user-attachments/assets/1dbbbcde-1557-4bd4-8541-78710f5f0842) 10. Check Windows Settings App -> Apps -> Startup (NOTICE HERE, Terminal is checked on and greyed out so it can't be disabled??) ![Image](https://github.com/user-attachments/assets/12235cc5-0a02-47d8-b026-2ac2045830ee) 11. On Fresh windows start, terminal auto opens and launches WSL Things I have tried to alleviate: 1. I have disabled WSL, which still causes Terminal to launch at startup, but will just launch with a WSL error because WSL is not running 2. I have turned off showing recent files in explorer 3. I have ensured Linux subsystem folders are not expanded or manually pinned to quick start 4. Ensured docker desktop with WSL support does not run at startup 5. https://www.winhelponline.com/blog/wp-content/uploads/2023/06/store_appx_startup_fix.zip registry hack which allows me to actually change Startup Apps under Settings App to disable Windows Terminal ### Expected Behavior I expect Windows Terminal to not autostart when computer starts, I also expect to be able to disable Windows Terminal from Settings App -> Apps -> Startup, currently it is enabled and greyed out with no way to disable it. ### Actual Behavior Windows terminal launches at startup everytime.
Issue-Bug,Needs-Triage
low
Critical
2,807,025,153
rust
Tracking issue for release notes of #126604: Uplift `clippy::double_neg` lint as `double_negations`
This issue tracks the release notes text for #126604. ### Steps - [ ] Proposed text is drafted by PR author (or team) making the noteworthy change. - [ ] Issue is nominated for release team review of clarity for wider audience. - [ ] Release team includes text in release notes/blog posts. ### Release notes text The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing). ````markdown # Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...) - [Uplift `clippy::double_neg` lint as `double_negations`](https://github.com/rust-lang/rust/pull/126604) ```` > [!TIP] > Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use. > The category will be de-duplicated with all the other ones by the release team. > > *More than one section can be included if needed.* ### Release blog section If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section. *Otherwise leave it empty.* ````markdown ```` cc @kadiwa4, @nnethercote -- origin issue/PR authors and assignees for starting to draft text
T-lang,relnotes,needs-triage,relnotes-tracking-issue
low
Minor
2,807,038,420
iptv
Edit: Workpoint 23 (1080p)
### Stream URL (required) http://appdootv2.dootvde.com:1935/live/50016_workpoint_tv.stream.smil/playlist.m3u8 ### Channel ID Workpoint23HD.th ### Channel Name Workpoint 23 ### Quality 1080p ### Label None ### Timeshift _No response_ ### HTTP User Agent _No response_ ### HTTP Referrer _No response_ ### Notes _No response_ ### Contributing Guide - [x] I have read [Contributing Guide](https://github.com/iptv-org/iptv/blob/master/CONTRIBUTING.md)
streams:edit
low
Minor
2,807,051,525
vscode
Provide way to disable Copilot Free signup prompt for enterprises
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> In many enterprises the free version of GitHub Copilot is disallowed for data security reasons. Therefore we do not want users to be prompted to sign up to Copilot Free when they start VSCode. We cannot work around this by removing the Copilot extension as we do allow Copilot enterprise, just not the free version.
workbench-copilot
low
Minor